Viewing a response to: @taskmaster4450le/re-taskmaster4450le-2e6ykpseq
The ever-shrinking model According to a study from researchers at Harvard, Stanford, MIT, Databricks, and Carnegie Mellon, quantized models perform worse if the original, unquantized version of the model was trained over a long period on lots of data. In other words, at a certain point, it may actually be better to just train a smaller model rather than cook down a big one. That could spell bad news for AI companies training extremely large models (known to improve answer quality) and then quantizing them in an effort to make them less expensive to serve.
author | taskmaster4450le |
---|---|
permlink | re-taskmaster4450le-9cvctje3 |
category | hive-167922 |
json_metadata | {"app":"leothreads/0.3","format":"markdown","tags":["leofinance"],"canonical_url":"https://inleo.io/threads/view/taskmaster4450le/re-taskmaster4450le-9cvctje3","isPoll":false,"pollOptions":{},"dimensions":[]} |
created | 2024-11-17 18:52:57 |
last_update | 2024-11-17 18:52:57 |
depth | 3 |
children | 0 |
last_payout | 2024-11-24 18:52:57 |
cashout_time | 1969-12-31 23:59:59 |
total_payout_value | 0.000 HBD |
curator_payout_value | 0.000 HBD |
pending_payout_value | 0.000 HBD |
promoted | 0.000 HBD |
body_length | 562 |
author_reputation | 2,193,078,876,199,132 |
root_title | "LeoThread 2024-11-17 10:12" |
beneficiaries | [] |
max_accepted_payout | 1,000,000.000 HBD |
percent_hbd | 10,000 |
post_id | 138,514,300 |
net_rshares | 0 |