Large language models like ChatGPT are producing falsehoods more accurately described as 'bullshit' rather than 'hallucinations'. These models generate human-like text by analyzing probabilities rather than aiming for truth. Describing their inaccuracies as bullshit is argued to be a more useful framework for understanding and
54 Comments
Sort: