-
Notifications
You must be signed in to change notification settings - Fork 1
/
Copy pathrun_2114.log
386 lines (385 loc) · 18.2 KB
/
run_2114.log
1
2
3
4
5
6
7
8
9
10
11
12
13
Python 3.12.5
['▁<', '|', 'endo', 'f', 'text', '|', '>']
[39, 19, 37, 35, 38, 19, 40]
['▁Before', '▁we', '▁pro', 'ce', 'ed', '▁any', '▁further', ',', '▁hear', '▁me', '▁speak', '.']
torch.Size([512, 128]) torch.Size([512, 128])
The <|endoftext|> id is : 1
total params: 12,597,504
model size: 48.243MB
Training on cuda.
Epoch 1/2:
Step 1/953240 - LR:0.0016 - train_loss: 8.462Step 11/953240 - LR:0.0016 - train_loss: 5.984Step 21/953240 - LR:0.0017 - train_loss: 5.952Step 31/953240 - LR:0.0016 - train_loss: 5.920Step 41/953240 - LR:0.0017 - train_loss: 5.908Step 51/953240 - LR:0.0016 - train_loss: 5.904Step 61/953240 - LR:0.0017 - train_loss: 5.906Step 71/953240 - LR:0.0016 - train_loss: 5.890Step 81/953240 - LR:0.0017 - train_loss: 5.784Step 91/953240 - LR:0.0016 - train_loss: 5.459Step 101/953240 - LR:0.0016 - train_loss: 5.235Step 111/953240 - LR:0.0016 - train_loss: 5.202Step 121/953240 - LR:0.0017 - train_loss: 5.137Step 131/953240 - LR:0.0016 - train_loss: 5.071Step 141/953240 - LR:0.0017 - train_loss: 5.044Step 151/953240 - LR:0.0016 - train_loss: 5.017Step 161/953240 - LR:0.0017 - train_loss: 4.957Step 171/953240 - LR:0.0016 - train_loss: 4.934Step 181/953240 - LR:0.0017 - train_loss: 4.910Step 191/953240 - LR:0.0016 - train_loss: 4.917Step 201/953240 - LR:0.0016 - train_loss: 4.789Step 211/953240 - LR:0.0016 - train_loss: 4.700Step 221/953240 - LR:0.0016 - train_loss: 4.665Step 231/953240 - LR:0.0016 - train_loss: 4.619Step 241/953240 - LR:0.0017 - train_loss: 4.536Step 251/953240 - LR:0.0016 - train_loss: 4.542Step 261/953240 - LR:0.0017 - train_loss: 4.426Step 271/953240 - LR:0.0016 - train_loss: 4.447Step 281/953240 - LR:0.0017 - train_loss: 4.341Step 291/953240 - LR:0.0016 - train_loss: 4.403Step 301/953240 - LR:0.0017 - train_loss: 4.269Step 311/953240 - LR:0.0016 - train_loss: 4.274Step 321/953240 - LR:0.0017 - train_loss: 4.179Step 331/953240 - LR:0.0016 - train_loss: 4.227Step 341/953240 - LR:0.0017 - train_loss: 4.093Step 351/953240 - LR:0.0016 - train_loss: 4.156Step 361/953240 - LR:0.0017 - train_loss: 4.088Step 371/953240 - LR:0.0016 - train_loss: 4.127Step 381/953240 - LR:0.0017 - train_loss: 3.998Step 391/953240 - LR:0.0016 - train_loss: 4.036Step 401/953240 - LR:0.0017 - train_loss: 3.971Step 411/953240 - LR:0.0016 - train_loss: 3.992Step 421/953240 - LR:0.0017 - train_loss: 3.924Step 431/953240 - LR:0.0016 - train_loss: 4.036Step 441/953240 - LR:0.0017 - train_loss: 3.950Step 451/953240 - LR:0.0016 - train_loss: 4.109Step 461/953240 - LR:0.0017 - train_loss: 3.919Step 471/953240 - LR:0.0016 - train_loss: 3.855Step 481/953240 - LR:0.0017 - train_loss: 3.747Step 491/953240 - LR:0.0016 - train_loss: 3.861Step 501/953240 - LR:0.0017 - train_loss: 3.803Step 511/953240 - LR:0.0016 - train_loss: 3.810Step 521/953240 - LR:0.0017 - train_loss: 3.690Step 531/953240 - LR:0.0016 - train_loss: 3.679Step 541/953240 - LR:0.0017 - train_loss: 3.664Step 551/953240 - LR:0.0016 - train_loss: 3.699Step 561/953240 - LR:0.0017 - train_loss: 3.638Step 571/953240 - LR:0.0016 - train_loss: 3.663Step 581/953240 - LR:0.0016 - train_loss: 3.626Step 591/953240 - LR:0.0016 - train_loss: 3.781Step 601/953240 - LR:0.0016 - train_loss: 3.670Step 611/953240 - LR:0.0016 - train_loss: 3.693Step 621/953240 - LR:0.0016 - train_loss: 3.626Step 631/953240 - LR:0.0016 - train_loss: 3.596Step 641/953240 - LR:0.0017 - train_loss: 3.544Step 651/953240 - LR:0.0016 - train_loss: 3.647Step 661/953240 - LR:0.0017 - train_loss: 3.588Step 671/953240 - LR:0.0016 - train_loss: 3.537Step 681/953240 - LR:0.0017 - train_loss: 3.552Step 691/953240 - LR:0.0016 - train_loss: 3.596Step 701/953240 - LR:0.0017 - train_loss: 3.495Step 711/953240 - LR:0.0016 - train_loss: 3.589Step 721/953240 - LR:0.0017 - train_loss: 3.526Step 731/953240 - LR:0.0016 - train_loss: 3.562Step 741/953240 - LR:0.0017 - train_loss: 3.467Step 751/953240 - LR:0.0016 - train_loss: 3.544Step 761/953240 - LR:0.0017 - train_loss: 3.476Step 771/953240 - LR:0.0016 - train_loss: 3.528Step 781/953240 - LR:0.0016 - train_loss: 3.430Step 791/953240 - LR:0.0016 - train_loss: 3.482Step 801/953240 - LR:0.0017 - train_loss: 3.377Step 811/953240 - LR:0.0016 - train_loss: 3.494Step 821/953240 - LR:0.0017 - train_loss: 3.429Step 831/953240 - LR:0.0016 - train_loss: 3.536Step 841/953240 - LR:0.0017 - train_loss: 3.449Step 851/953240 - LR:0.0016 - train_loss: 3.500Step 861/953240 - LR:0.0017 - train_loss: 3.390Step 871/953240 - LR:0.0016 - train_loss: 3.401Step 881/953240 - LR:0.0017 - train_loss: 3.376Step 891/953240 - LR:0.0016 - train_loss: 3.470Step 901/953240 - LR:0.0017 - train_loss: 3.340Step 911/953240 - LR:0.0016 - train_loss: 3.402Step 921/953240 - LR:0.0017 - train_loss: 3.342Step 931/953240 - LR:0.0016 - train_loss: 3.364Step 941/953240 - LR:0.0017 - train_loss: 3.308Step 951/953240 - LR:0.0016 - train_loss: 3.287Step 961/953240 - LR:0.0017 - train_loss: 3.276Step 971/953240 - LR:0.0016 - train_loss: 3.294Step 981/953240 - LR:0.0017 - train_loss: 3.223Step 991/953240 - LR:0.0016 - train_loss: 3.373Step 1001/953240 - LR:0.0017 - train_loss: 3.314Step 1011/953240 - LR:0.0016 - train_loss: 3.414Step 1021/953240 - LR:0.0017 - train_loss: 3.264Step 1031/953240 - LR:0.0016 - train_loss: 3.338Step 1041/953240 - LR:0.0017 - train_loss: 3.266Step 1051/953240 - LR:0.0016 - train_loss: 3.275Step 1061/953240 - LR:0.0017 - train_loss: 3.261Step 1071/953240 - LR:0.0016 - train_loss: 3.269Step 1081/953240 - LR:0.0017 - train_loss: 3.221Step 1091/953240 - LR:0.0016 - train_loss: 3.283Step 1101/953240 - LR:0.0017 - train_loss: 3.216Step 1111/953240 - LR:0.0016 - train_loss: 3.295Step 1121/953240 - LR:0.0017 - train_loss: 3.225Step 1131/953240 - LR:0.0016 - train_loss: 3.235Step 1141/953240 - LR:0.0017 - train_loss: 3.165Step 1151/953240 - LR:0.0016 - train_loss: 3.210Step 1161/953240 - LR:0.0017 - train_loss: 3.158Step 1171/953240 - LR:0.0016 - train_loss: 3.305Step 1181/953240 - LR:0.0017 - train_loss: 3.166Step 1191/953240 - LR:0.0016 - train_loss: 3.280Step 1201/953240 - LR:0.0017 - train_loss: 3.225Step 1211/953240 - LR:0.0016 - train_loss: 3.180Step 1221/953240 - LR:0.0017 - train_loss: 3.109Step 1231/953240 - LR:0.0016 - train_loss: 3.174Step 1241/953240 - LR:0.0017 - train_loss: 3.112Step 1251/953240 - LR:0.0016 - train_loss: 3.238Step 1261/953240 - LR:0.0017 - train_loss: 3.115Step 1271/953240 - LR:0.0016 - train_loss: 3.156Step 1281/953240 - LR:0.0017 - train_loss: 3.079Step 1291/953240 - LR:0.0016 - train_loss: 3.169Step 1301/953240 - LR:0.0017 - train_loss: 3.115Step 1311/953240 - LR:0.0016 - train_loss: 3.100Step 1321/953240 - LR:0.0017 - train_loss: 3.041Step 1331/953240 - LR:0.0016 - train_loss: 3.155Step 1341/953240 - LR:0.0016 - train_loss: 3.071Step 1351/953240 - LR:0.0016 - train_loss: 3.146Step 1361/953240 - LR:0.0017 - train_loss: 3.060Step 1371/953240 - LR:0.0016 - train_loss: 3.143Step 1381/953240 - LR:0.0017 - train_loss: 3.009Step 1391/953240 - LR:0.0016 - train_loss: 3.042Step 1401/953240 - LR:0.0017 - train_loss: 2.996Step 1411/953240 - LR:0.0016 - train_loss: 3.081Step 1421/953240 - LR:0.0016 - train_loss: 3.043Step 1431/953240 - LR:0.0016 - train_loss: 3.011Step 1441/953240 - LR:0.0017 - train_loss: 3.014Step 1451/953240 - LR:0.0016 - train_loss: 3.082Step 1461/953240 - LR:0.0017 - train_loss: 2.996Step 1471/953240 - LR:0.0016 - train_loss: 3.006Step 1481/953240 - LR:0.0017 - train_loss: 2.981Step 1491/953240 - LR:0.0016 - train_loss: 3.040Step 1501/953240 - LR:0.0017 - train_loss: 2.936Step 1511/953240 - LR:0.0016 - train_loss: 2.997Step 1521/953240 - LR:0.0017 - train_loss: 2.968Step 1531/953240 - LR:0.0016 - train_loss: 3.055Step 1541/953240 - LR:0.0017 - train_loss: 2.969Step 1551/953240 - LR:0.0016 - train_loss: 3.013Step 1561/953240 - LR:0.0016 - train_loss: 2.928Step 1571/953240 - LR:0.0016 - train_loss: 3.030Step 1581/953240 - LR:0.0017 - train_loss: 2.928Step 1591/953240 - LR:0.0016 - train_loss: 2.956Step 1601/953240 - LR:0.0017 - train_loss: 2.877Step 1611/953240 - LR:0.0016 - train_loss: 2.939Step 1621/953240 - LR:0.0017 - train_loss: 2.888Step 1631/953240 - LR:0.0016 - train_loss: 2.927Step 1641/953240 - LR:0.0016 - train_loss: 2.880Step 1651/953240 - LR:0.0016 - train_loss: 2.899Step 1661/953240 - LR:0.0017 - train_loss: 2.838Step 1671/953240 - LR:0.0016 - train_loss: 2.956Step 1681/953240 - LR:0.0017 - train_loss: 2.873Step 1691/953240 - LR:0.0016 - train_loss: 2.906Step 1701/953240 - LR:0.0017 - train_loss: 2.880Step 1711/953240 - LR:0.0016 - train_loss: 2.917Step 1721/953240 - LR:0.0016 - train_loss: 2.811Step 1731/953240 - LR:0.0016 - train_loss: 2.876Step 1741/953240 - LR:0.0017 - train_loss: 2.807Step 1751/953240 - LR:0.0016 - train_loss: 2.899Step 1761/953240 - LR:0.0017 - train_loss: 2.844Step 1771/953240 - LR:0.0016 - train_loss: 2.882Step 1781/953240 - LR:0.0017 - train_loss: 2.814Step 1791/953240 - LR:0.0016 - train_loss: 2.875Step 1801/953240 - LR:0.0017 - train_loss: 2.807Step 1811/953240 - LR:0.0016 - train_loss: 2.806Step 1821/953240 - LR:0.0017 - train_loss: 2.799Step 1831/953240 - LR:0.0016 - train_loss: 2.834Step 1841/953240 - LR:0.0017 - train_loss: 2.765Step 1851/953240 - LR:0.0016 - train_loss: 2.816Step 1861/953240 - LR:0.0017 - train_loss: 2.782Step 1871/953240 - LR:0.0016 - train_loss: 2.836Step 1881/953240 - LR:0.0017 - train_loss: 2.824Step 1891/953240 - LR:0.0016 - train_loss: 2.826Step 1901/953240 - LR:0.0017 - train_loss: 2.780Step 1911/953240 - LR:0.0016 - train_loss: 2.812Step 1921/953240 - LR:0.0017 - train_loss: 2.775Step 1931/953240 - LR:0.0016 - train_loss: 2.765Step 1941/953240 - LR:0.0016 - train_loss: 2.755Step 1951/953240 - LR:0.0016 - train_loss: 2.772Step 1961/953240 - LR:0.0017 - train_loss: 2.689Step 1971/953240 - LR:0.0016 - train_loss: 2.753Step 1981/953240 - LR:0.0017 - train_loss: 2.757Step 1991/953240 - LR:0.0016 - train_loss: 2.750Step 2001/953240 - LR:0.0017 - train_loss: 2.744Step 2011/953240 - LR:0.0016 - train_loss: 2.734Step 2021/953240 - LR:0.0016 - train_loss: 2.732Step 2031/953240 - LR:0.0016 - train_loss: 2.739Step 2041/953240 - LR:0.0017 - train_loss: 2.671Step 2051/953240 - LR:0.0016 - train_loss: 2.711Step 2061/953240 - LR:0.0017 - train_loss: 2.718Step 2071/953240 - LR:0.0016 - train_loss: 2.728Step 2081/953240 - LR:0.0017 - train_loss: 2.716Step 2091/953240 - LR:0.0016 - train_loss: 2.759Step 2101/953240 - LR:0.0017 - train_loss: 2.678Step 2111/953240 - LR:0.0016 - train_loss: 2.730Step 2121/953240 - LR:0.0017 - train_loss: 2.664Step 2131/953240 - LR:0.0016 - train_loss: 2.685Step 2141/953240 - LR:0.0017 - train_loss: 2.643Step 2151/953240 - LR:0.0016 - train_loss: 2.669Step 2161/953240 - LR:0.0017 - train_loss: 2.642Step 2171/953240 - LR:0.0016 - train_loss: 2.745Step 2181/953240 - LR:0.0017 - train_loss: 2.668Step 2191/953240 - LR:0.0016 - train_loss: 2.708Step 2201/953240 - LR:0.0017 - train_loss: 2.610Step 2211/953240 - LR:0.0016 - train_loss: 2.712Step 2221/953240 - LR:0.0017 - train_loss: 2.623Step 2231/953240 - LR:0.0016 - train_loss: 2.665Step 2241/953240 - LR:0.0017 - train_loss: 2.595Step 2251/953240 - LR:0.0016 - train_loss: 2.643Step 2261/953240 - LR:0.0017 - train_loss: 2.594Step 2271/953240 - LR:0.0016 - train_loss: 2.618Step 2281/953240 - LR:0.0017 - train_loss: 2.570Step 2291/953240 - LR:0.0016 - train_loss: 2.621Step 2301/953240 - LR:0.0017 - train_loss: 2.554Step 2311/953240 - LR:0.0016 - train_loss: 2.596Step 2321/953240 - LR:0.0016 - train_loss: 2.582Step 2331/953240 - LR:0.0016 - train_loss: 2.622Step 2341/953240 - LR:0.0017 - train_loss: 2.541Step 2351/953240 - LR:0.0016 - train_loss: 2.637Step 2361/953240 - LR:0.0017 - train_loss: 2.583Step 2371/953240 - LR:0.0016 - train_loss: 2.675Step 2381/953240 - LR:0.0017 - train_loss: 2.552Step 2391/953240 - LR:0.0016 - train_loss: 2.636Step 2401/953240 - LR:0.0017 - train_loss: 2.543Step 2411/953240 - LR:0.0016 - train_loss: 2.609Step 2421/953240 - LR:0.0017 - train_loss: 2.522Step 2431/953240 - LR:0.0016 - train_loss: 2.587Step 2441/953240 - LR:0.0017 - train_loss: 2.541Step 2451/953240 - LR:0.0016 - train_loss: 2.602Step 2461/953240 - LR:0.0017 - train_loss: 2.534Step 2471/953240 - LR:0.0016 - train_loss: 2.604Step 2481/953240 - LR:0.0017 - train_loss: 2.525Step 2491/953240 - LR:0.0016 - train_loss: 2.556Step 2501/953240 - LR:0.0017 - train_loss: 2.526Step 2511/953240 - LR:0.0016 - train_loss: 2.556Step 2521/953240 - LR:0.0017 - train_loss: 2.549Step 2531/953240 - LR:0.0016 - train_loss: 2.575Step 2541/953240 - LR:0.0017 - train_loss: 2.472Step 2551/953240 - LR:0.0016 - train_loss: 2.560Step 2561/953240 - LR:0.0017 - train_loss: 2.509Step 2571/953240 - LR:0.0016 - train_loss: 2.521Step 2581/953240 - LR:0.0017 - train_loss: 2.466Step 2591/953240 - LR:0.0016 - train_loss: 2.518Step 2601/953240 - LR:0.0017 - train_loss: 2.468Step 2611/953240 - LR:0.0016 - train_loss: 2.533Step 2621/953240 - LR:0.0017 - train_loss: 2.448Step 2631/953240 - LR:0.0016 - train_loss: 2.501Step 2641/953240 - LR:0.0017 - train_loss: 2.479Step 2651/953240 - LR:0.0016 - train_loss: 2.491Step 2661/953240 - LR:0.0017 - train_loss: 2.471Step 2671/953240 - LR:0.0016 - train_loss: 2.512Step 2681/953240 - LR:0.0017 - train_loss: 2.447Step 2691/953240 - LR:0.0016 - train_loss: 2.517Step 2701/953240 - LR:0.0017 - train_loss: 2.486Step 2711/953240 - LR:0.0016 - train_loss: 2.483Step 2721/953240 - LR:0.0017 - train_loss: 2.433Step 2731/953240 - LR:0.0016 - train_loss: 2.526Step 2741/953240 - LR:0.0017 - train_loss: 2.439Step 2751/953240 - LR:0.0016 - train_loss: 2.449Step 2761/953240 - LR:0.0017 - train_loss: 2.443Step 2771/953240 - LR:0.0016 - train_loss: 2.524Step 2781/953240 - LR:0.0017 - train_loss: 2.449Step 2791/953240 - LR:0.0016 - train_loss: 2.485Step 2801/953240 - LR:0.0017 - train_loss: 2.465Step 2811/953240 - LR:0.0016 - train_loss: 2.464Step 2821/953240 - LR:0.0016 - train_loss: 2.413Step 2831/953240 - LR:0.0016 - train_loss: 2.460Step 2841/953240 - LR:0.0017 - train_loss: 2.407Step 2851/953240 - LR:0.0016 - train_loss: 2.455Step 2861/953240 - LR:0.0017 - train_loss: 2.408Step 2871/953240 - LR:0.0016 - train_loss: 2.416Step 2881/953240 - LR:0.0017 - train_loss: 2.404Step 2891/953240 - LR:0.0016 - train_loss: 2.465Step 2901/953240 - LR:0.0017 - train_loss: 2.412Step 2911/953240 - LR:0.0016 - train_loss: 2.483Step 2921/953240 - LR:0.0017 - train_loss: 2.384Step 2931/953240 - LR:0.0016 - train_loss: 2.456Step 2941/953240 - LR:0.0017 - train_loss: 2.358Step 2951/953240 - LR:0.0016 - train_loss: 2.428Step 2961/953240 - LR:0.0017 - train_loss: 2.378Step 2971/953240 - LR:0.0016 - train_loss: 2.457Step 2981/953240 - LR:0.0016 - train_loss: 2.388Step 2991/953240 - LR:0.0016 - train_loss: 2.437Step 3001/953240 - LR:0.0017 - train_loss: 2.408Step 3011/953240 - LR:0.0016 - train_loss: 2.441Step 3021/953240 - LR:0.0017 - train_loss: 2.358Step 3031/953240 - LR:0.0016 - train_loss: 2.414Step 3041/953240 - LR:0.0017 - train_loss: 2.381Step 3051/953240 - LR:0.0016 - train_loss: 2.414Step 3061/953240 - LR:0.0017 - train_loss: 2.370Step 3071/953240 - LR:0.0016 - train_loss: 2.384Step 3081/953240 - LR:0.0017 - train_loss: 2.339Step 3091/953240 - LR:0.0016 - train_loss: 2.380Step 3101/953240 - LR:0.0017 - train_loss: 2.341Step 3111/953240 - LR:0.0016 - train_loss: 2.366Step 3121/953240 - LR:0.0017 - train_loss: 2.307Step 3131/953240 - LR:0.0016 - train_loss: 2.371Step 3141/953240 - LR:0.0017 - train_loss: 2.324Step 3151/953240 - LR:0.0016 - train_loss: 2.354Step 3161/953240 - LR:0.0017 - train_loss: 2.355Step 3171/953240 - LR:0.0016 - train_loss: 2.353Step 3181/953240 - LR:0.0017 - train_loss: 2.319Step 3191/953240 - LR:0.0016 - train_loss: 2.368Step 3201/953240 - LR:0.0016 - train_loss: 2.322Step 3211/953240 - LR:0.0016 - train_loss: 2.356Step 3221/953240 - LR:0.0017 - train_loss: 2.299Step 3231/953240 - LR:0.0016 - train_loss: 2.342Step 3241/953240 - LR:0.0017 - train_loss: 2.326Step 3251/953240 - LR:0.0016 - train_loss: 2.356Step 3261/953240 - LR:0.0017 - train_loss: 2.304Step 3271/953240 - LR:0.0016 - train_loss: 2.307Step 3281/953240 - LR:0.0017 - train_loss: 2.341Step 3291/953240 - LR:0.0016 - train_loss: 2.307Step 3301/953240 - LR:0.0017 - train_loss: 2.302Step 3311/953240 - LR:0.0016 - train_loss: 2.353Step 3321/953240 - LR:0.0017 - train_loss: 2.301Step 3331/953240 - LR:0.0016 - train_loss: 2.362Step 3341/953240 - LR:0.0017 - train_loss: 2.294Step 3351/953240 - LR:0.0016 - train_loss: 2.321Step 3361/953240 - LR:0.0016 - train_loss: 2.270Step 3371/953240 - LR:0.0016 - train_loss: 2.355Step 3381/953240 - LR:0.0017 - train_loss: 2.274Step 3391/953240 - LR:0.0016 - train_loss: 2.317Step 3401/953240 - LR:0.0017 - train_loss: 2.311Step 3411/953240 - LR:0.0016 - train_loss: 2.315Step 3421/953240 - LR:0.0017 - train_loss: 2.297Step 3431/953240 - LR:0.0016 - train_loss: 2.341Step 3441/953240 - LR:0.0017 - train_loss: 2.303Step 3451/953240 - LR:0.0016 - train_loss: 2.304Step 3461/953240 - LR:0.0017 - train_loss: 2.255Step 3471/953240 - LR:0.0016 - train_loss: 2.315Step 3481/953240 - LR:0.0017 - train_loss: 2.221Step 3491/953240 - LR:0.0016 - train_loss: 2.315Step 3501/953240 - LR:0.0017 - train_loss: 2.260Step 3511/953240 - LR:0.0016 - train_loss: 2.351Step 3521/953240 - LR:0.0017 - train_loss: 2.240Step 3531/953240 - LR:0.0016 - train_loss: 2.322Step 3541/953240 - LR:0.0017 - train_loss: 2.251Step 3551/953240 - LR:0.0016 - train_loss: 2.283Step 3561/953240 - LR:0.0017 - train_loss: 2.275Step 3571/953240 - LR:0.0016 - train_loss: 2.287Step 3581/953240 - LR:0.0016 - train_loss: 2.215Step 3591/953240 - LR:0.0016 - train_loss: 2.285Step 3601/953240 - LR:0.0017 - train_loss: 2.220Step 3611/953240 - LR:0.0016 - train_loss: 2.293Step 3621/953240 - LR:0.0017 - train_loss: 2.248Step 3631/953240 - LR:0.0016 - train_loss: 2.273Step 3641/953240 - LR:0.0017 - train_loss: 2.209Step 3651/953240 - LR:0.0016 - train_loss: 2.261Step 3661/953240 - LR:0.0017 - train_loss: 2.281Step 3671/953240 - LR:0.0016 - train_loss: 2.253Step 3681/953240 - LR:0.0017 - train_loss: 2.226Step 3691/953240 - LR:0.0016 - train_loss: 2.246Step 3701/953240 - LR:0.0017 - train_loss: 2.225Step 3711/953240 - LR:0.0016 - train_loss: 2.262slurmstepd: error: *** STEP 55301.0 ON makkapakka02 CANCELLED AT 2024-10-24T21:50:11 ***
Executing command >>>>
srun --pty -c 10 -p makkapakka --export=ALL --gpus=4 ./run.sh