Summary of "“What's wrong with LLMs and what we should be building instead” - Tom Dietterich - #VSCF2023"

Talk overview

Key capabilities acknowledged

LLMs demonstrate important strengths:

Main problems (with examples)

  1. Incorrect and inconsistent outputs (hallucination)

    • LLMs invent facts or contradict themselves.

      Example: A GPT-2 story that says a unicorn both has one horn and four horns.

    • TruthfulQA benchmark: earlier models perform poorly; GPT-4 (with special training) barely exceeds 50% on hard truthfulness queries.

    • Fabricated citations, people, or events have been reported (invented journal articles, false accusations).
  2. Dangerous / biased / socially unacceptable outputs

    • Models can produce racist/sexist content or dangerous instructions.

      Example (transcript): a generated Python function that uses race/gender to decide who is a “good scientist”.

    • Prompting-workarounds and storytelling can be used to elicit harmful outputs despite safety training.

  3. Expensive to train and hard to update

    • Very large training costs (GPT-4 quoted > $100M).
    • Knowledge encoded in weights becomes stale and is expensive to correct — LLMs support “ask” but not an inexpensive “tell”.
  4. Lack of attribution and provenance

    • It is difficult to trace which training documents produced a particular output.
    • Retrieval-augmented systems can add links, but attribution is often incorrect or unsupported.
  5. Poor non-linguistic and structured reasoning

    • Weak spatial/mental-model reasoning (e.g., incorrect answers about positions in a room).
    • Weak formal reasoning, planning, and process reasoning (actions, preconditions, side effects).
  6. Miscalibration and overconfidence

    • Techniques like RLHF reduce certain harms but can damage probability calibration: models become overconfident and less likely to say “I don’t know.”
  7. Security / data-poisoning risks

    • Web retrieval exposes models to adversarial documents that can inject harmful instructions or manipulate behavior.

Diagnoses / conceptual framing

Existing mitigation strategies (and limitations)

Proposed direction (Dietterich’s recommendations)

  1. Move to modular architectures inspired by cognitive neuroscience

    • Separate components for language (syntax/semantics), factual/world knowledge (updatable KB), commonsense reasoning, episodic memory, situation models, formal reasoning/planning, and a metacognitive (prefrontal-like) orchestrator.
  2. Use structured knowledge representations (knowledge graphs)

    • Extract facts from text into a KB; infer communicative goals/pragmatics.
    • Design an encoder that maps paragraph → detected facts + communicative intent and adds facts to the KB with evidence accumulation.
    • Design a decoder that generates text from explicit KB facts and goals, enabling explicit fact extraction and attribution as side-effects.
    • Revisit projects like NELL (Never-Ending Language Learning), bootstrapped with modern LLMs to populate and maintain KBs.
  3. Train systems to return answers plus arguments / justifications

    • Require explicit argumentation and provenance so downstream consumers and modules can evaluate soundness.
    • Use formal argumentation and knowledge-representation techniques to handle inconsistency, multiple viewpoints, and cultural differences (micro-worlds approach).
  4. Integrate reasoning and planning more tightly

    • Make formal reasoning, proof assistants, and planners native parts of the system or tightly coordinated subsystems rather than mere external “tools.”
  5. Improve calibration and the ability to say “I don’t know”

    • Develop competence estimation and refusal mechanisms; estimate epistemic uncertainty separately from generative confidence.
  6. Prioritize inspectability and provenance

    • Make sources, preference models, and safety rules auditable to address bias and contested norms.
  7. Support research infrastructure and openness

    • Governments and funders should provide compute/resources for academia and smaller groups to experiment openly.
    • Open-source LLM efforts (e.g., Alpaca activity) accelerate research and fixes.

Practical application notes

Unresolved research challenges

Cited / referenced works, systems and actors

Bottom line

Dietterich argues that LLMs are powerful but flawed because they are monolithic statistical models rather than modular AI systems with explicit, updatable knowledge, provenance, reasoning, and metacognition. He proposes a research program to build modular architectures (knowledge graphs, episodic memory, planners, argumentation), integrate verification tools, and produce systems that justify claims, can be efficiently updated, and are auditable.

Category ?

Technology


Share this summary


Is the summary off?

If you think the summary is inaccurate, you can reprocess it with the latest model.

Video