TheChurn@kbin.socialtoTechnology@beehaw.org•Hallucination is Inevitable: An Innate Limitation of Large Language Models (arxiv preprint)
0·
8 months agoA token is not a concept. A token is a word or word fragment that occured often in free text and was assigned a number. Common words, prefixes, and suffixes are the vast majority of tokens, and the rest are uncommon pairs of letters.
The algorithm to generate tokens is essentially compression, there is no semantic meaning embedded in them.
One thing that the article didn’t touch on, since it was focused on input costs, is the extra pollution from using EVs.
EVs are substantially heavier than ICEs of the same class, due to the battery. This leads to extra wear on the tires, break pads, and road surface -> even more micro plastics and particulate air pollution.
We need to reduce our ecological footprint, not merely change it from oil to reactive metals.