Aligning Generalization Between Humans and Machines

Aligning Generalization Between Humans and Machines

Notes from “Aligning generalization between humans and machines”.

“Generalization” is the ability to use what you learned before to handle new situations. In human–AI collaboration, generalization matters because the AI might face cases that look different from its training data, while humans often expect it to adapt in a sensible way.

Why humans and machines generalize differently

Humans often build compact concepts (“rules of thumb”) and can reuse them in new contexts through analogy (e.g., applying an old idea to a new domain). Machines—especially modern statistical AI—often learn patterns from large datasets and represent knowledge as probability-like behaviors. This can work extremely well on familiar data, but it may fail when the new task or domain is far from what the model has seen.

A helpful way to think about “generalization”

A clear framework is to treat generalization in three ways:

  • Generalization as a process: how a system learns from experience (e.g., abstraction or learning from data).
  • Generalization as a product: what is learned (rules, concepts, prototypes, exemplars, or probability distributions).
  • Generalization as an operator: applying what was learned to new inputs and tasks successfully.

Humans and machines can differ at all three levels, which helps explain why they sometimes disagree or fail to collaborate smoothly.

Three broad machine approaches (and what they’re good at)

Many AI methods can be grouped by how they connect data (“instances”) to broader models (“principles”):

  • Statistical methods (deep learning, large neural models): strong at accuracy and scaling, but often weak at out-of-distribution robustness and transparent reasoning.
  • Knowledge-informed / symbolic methods (rules, logic, causal models): strong at compositional reasoning and interpretability, but harder to scale and harder to build for messy real-world data.
  • Instance-based methods (nearest neighbors, case-based reasoning, retrieval): strong at handling distribution shifts and remembering examples, but highly dependent on good representations and context.

A big idea in current research is to combine these strengths, for example through neurosymbolic methods.

Why evaluation is hard in the foundation-model era

With large “foundation models,” evaluation can be misleading:

  • Test sets may be contaminated (the model may have seen similar data during training).
  • Benchmarks can reward shortcuts (spurious correlations).
  • Public leaderboards can drive overfitting to the benchmark rather than real-world generalization.

That’s why researchers are increasingly interested in tests of abstraction, analogy, compositionality, robustness under shifts, and more careful evaluation setups.

What this means for future human–AI teams

Good teamwork needs more than “high accuracy.” When humans and AI disagree (for example in medical decisions), the team needs ways to:

  • detect misalignment,
  • explain both sides in shared concepts,
  • correct errors and realign the process—not just the final output.

In short, aligning generalization is about making AI behave in ways that match human expectations, and building tools and evaluations that reveal when it doesn’t.

Good Expressions To Notes

Generalization theory in ML is limited in several ways. It typically predicts that generalization only occurs if the available data are large enough to not just be memorized. By contrast, humans can generalize from a few samples for a specific task, as generalization in humans is not a singular event but based on lifelong experience of regularities observed in nature. Few-shot learning addresses this to some extent.

Gated Attention: A Simple Fix for Softmax Attention

Gated Attention: A Simple Fix for Softmax Attention

Notes from “Gated Attention for Large Language Models”, showing how a simple post-attention sigmoid gate improves performance, training stability, and long-context generalization by introducing non-linearity and sparsity while removing attention sinks.

This post summarizes key takeaways from Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-Free (arXiv:2505.06708v1, May 2025).

Academia Roadmap

An interactive publications timeline showcasing my academia roadmap, including published papers, under-review work, and ongoing projects across NLP, time-series modeling, and remote sensing.

Publications timeline

In preparation
Work in progress on grammar-knowledge-guided RL for reasoning-oriented LLMs.
In preparation
Diversification pipeline and score dynamics to evaluate robustness and generalization.
Long paper · ACL 2025 (Industry Track) · Published
Li, L., Sleem, L., Gentile, N., Nichil, G., & State, R.
Long paper · Procedia Computer Science 264, 242–251, 2025 · Published
Li, L., Sleem, L., Gentile, N., Nichil, G., & State, R.
Short paper · AAAI 2026 AI4TS & NeurIPS 2025 BERT2S · Accepted
Li, L., Sleem, L., Wang, Y, Xu, Y., Gentile, N., & State, R.
Long paper · IEEE CCNC 2026 · Accepted (arXiv:2511.11784)
Sleem, L., François, J., Li, L., Foucher, N., Gentile, N., & State, R.
Short paper · IEEE IGARSS 2025 · Published
Li, L., Wang, Y., & State, R.
Long paper · Under review at LoResMT 2026 @ EACL
Song, Y., Li, L., et al.
Long paper · Under review at ISPRS 2026 (XXV ISPRS Congress)
Wang, Y., Li, L., Yue, M., & State, R.

Pagination