The author discusses the limitations and potential risks of Large Language Models (LLMs), which are prone to confabulation, lying, and making up nonsense, and can be unpredictable and misleading. Despite their impressive capabilities, LLMs are far from human-equivalent intelligence and can only excel in specific, narrow tasks, making it difficult to predict their suitability for a given task.