Skip to content

cybertronai/schmidhuber-problems

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

1 Commit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

schmidhuber-problems

Stubs for the synthetic learning problems and toy datasets that appear in Jürgen Schmidhuber's papers (and those of his students/collaborators) from his 1987 diploma thesis through 2025.

Companion to hinton-problems: same scaffold style, different lineage. Each problem lives in its own folder containing:

  • README.md — source paper, brief description, what it demonstrates
  • problem.py — skeleton for dataset generation + model + training (raises NotImplementedError)

The catalog focuses on synthetic toy problems Schmidhuber (or close collaborators) designed to isolate an algorithmic capability. The signature Schmidhuber-style toy is the long-time-lag temporal indexing task — from the 1990 flip-flop through the 1992 chunker's 22-symbol task to the 1997 LSTM benchmark suite. Folders are flat; the catalog below is grouped by year for readability.

Visual tour: VISUAL_TOUR.md is the picture-first walk through the same catalog (skeletal until problems get implemented; structured to grow as GIFs and viz folders are added per stub).

Catalog

1980s — Local rules and the Neural Bucket Brigade

Schmidhuber (1989) — A local learning algorithm for dynamic feedforward and recurrent networks

1990 — Controller + world-model + flip-flop

Schmidhuber (1990) — Making the world differentiable (FKI-126-90)

  • flip-flop — output 1 iff B follows A with arbitrary delay (the canonical LSTM-precursor latch)
  • pole-balance-non-markov — cart-pole with hidden velocities, perfect differentiable model

Schmidhuber (1990) — Recurrent networks adjusted by adaptive critics

Schmidhuber & Huber (1990) — Learning to generate focus trajectories (FKI-128-90)

1991 — Curiosity, subgoals, the chunker

Schmidhuber (1991) — Adaptive confidence and adaptive curiosity (FKI-149-91)

Schmidhuber (1991) — Learning to generate sub-goals for action sequences (ICANN-91)

Schmidhuber (1991) — Reinforcement learning in Markovian and non-Markovian environments (NIPS-3)

  • pomdp-flag-maze — recurrent model+controller disambiguates hidden state

Schmidhuber (1991/1992) — Neural sequence chunkers / Learning complex extended sequences using the principle of history compression

1992 — Neural Computation triple

Schmidhuber (1992) — Learning to control fast-weight memories (NC 4(1))

Schmidhuber (1992) — Learning factorial codes by predictability minimization (NC 4(6))

1993 — Predictable classifications, self-reference, very deep chunking

Schmidhuber & Prelinger (1993) — Discovering predictable classifications (NC 5(4))

Schmidhuber (1993) — A self-referential weight matrix (ICANN-93)

Schmidhuber (1993) — Habilitationsschrift, Netzwerkarchitekturen, Zielfunktionen und Kettenregel

1995–1997 — Levin search and the LSTM benchmark suite

Schmidhuber (1995/1997) — Discovering solutions with low Kolmogorov complexity (ICML / NN 10)

Hochreiter & Schmidhuber (1996) — LSTM can solve hard long time lag problems (NIPS 9)

  • rs-two-sequence — random-weight-guessing breaks the Bengio-94 latch
  • rs-parity — random-weight-guessing on long-sequence parity
  • rs-tomita — RS attacks Tomita grammars #1, #2, #4
  • adding-problem — first non-trivial LSTM benchmark (Experiment 4)

Hochreiter & Schmidhuber (1997) — Long Short-Term Memory (NC 9(8)) — the canonical 6-experiment battery

Mid-90s — Evolutionary, RL, and feature detection

Salustowicz & Schmidhuber (1997) — Probabilistic Incremental Program Evolution

Schmidhuber, Zhao, Wiering (1997) — Shifting inductive bias with SSA (ML 28)

Wiering & Schmidhuber (1997) — HQ-learning (Adaptive Behavior 6(2))

Schmidhuber, Eldracher, Foltin (1996) — Semilinear PM produces well-known feature detectors (NC 8(4))

Hochreiter & Schmidhuber (1999) — Feature extraction through LOCOCODE (NC 11)

  • lococode-ica — flat-minimum search produces ICA-like sparse codes

2000–2002 — LSTM follow-ups

Gers, Schmidhuber, Cummins (2000) — Learning to forget (NC 12(10))

Gers & Schmidhuber (2001) — Context-free and context-sensitive languages (IEEE TNN 12(6))

Gers, Schraudolph, Schmidhuber (2002) — Learning precise timing (JMLR 3)

Eck & Schmidhuber (2002) — Blues improvisation with LSTM (NNSP)

2002–2010 — Evolutionary RL, OOPS, BLSTM+CTC

Schmidhuber, Wierstra, Gomez (2005/2007) — Evolino

Gomez & Schmidhuber (2005) — Co-evolving recurrent neurons (GECCO)

Graves et al. (2005/2006) — BLSTM and Connectionist Temporal Classification

Graves, Liwicki, Fernández, Bertolami, Bunke, Schmidhuber (2009) — Unconstrained handwriting (TPAMI)

Schmidhuber (2002–2004) — Optimal Ordered Problem Solver (ML 54)

2010–2017 — Deep learning at scale

Cireşan, Meier, Gambardella, Schmidhuber (2010) — Deep, big, simple nets (NC 22(12))

Cireşan, Meier, Schmidhuber (2012) — Multi-column deep neural networks (CVPR)

Cireşan, Giusti, Gambardella, Schmidhuber (2012) — EM segmentation (NIPS)

Srivastava, Masci, Kazerounian, Gomez, Schmidhuber (2013) — Compete to compute (NIPS)

Srivastava, Greff, Schmidhuber (2015) — Training very deep networks (NIPS)

Greff, Srivastava, Koutník, Steunebrink, Schmidhuber (2017) — LSTM: a search space odyssey (TNNLS)

Koutník, Greff, Gomez, Schmidhuber (2014) — A clockwork RNN (ICML)

  • clockwork-rnn — multi-rate hidden modules; audio gen, raw-audio TIMIT word

Koutník, Cuccu, Schmidhuber, Gomez (2013) — Vision-based RL via evolution (GECCO)

Greff, van Steenkiste, Schmidhuber (2017) — Neural Expectation Maximization (NIPS)

van Steenkiste, Chang, Greff, Schmidhuber (2018) — Relational Neural EM (ICLR)

2018–2025 — World models, fast-weight Transformers, systematic generalization

Ha & Schmidhuber (2018) — Recurrent World Models Facilitate Policy Evolution (NeurIPS)

Schmidhuber et al. (2019) — Reinforcement Learning Upside Down (arXiv)

  • upside-down-rl — RL as supervised learning conditioned on desired return

Schlag, Irie, Schmidhuber (2021) — Linear Transformers are secretly fast weight programmers (ICML)

Csordás, Irie, Schmidhuber (2022) — The Neural Data Router (ICLR)

Structure

problem-folder/
├── README.md      one paragraph: source + property
└── problem.py     stubs: generate_dataset, build_model, train

The stubs raise NotImplementedError. Fill in the parts you need.

Methodological caveat

Many of the early TUM technical-report PDFs (FKI-125-90, FKI-129-90, FKI-148-91, FKI-149-91, the 1993 Habilitationsschrift, Hochreiter's 1991 diploma thesis) are difficult to retrieve in original form. Stub READMEs reconstruct the experiments from corroborated secondary sources — Schmidhuber's Deep Learning: Our Miraculous Year 1990–1991 (2020), the 1997 LSTM paper's literature review, the 2001 Hochreiter/Bengio/Frasconi/Schmidhuber chapter Gradient Flow in Recurrent Nets, and relevant NeurIPS/Springer abstracts — and flag claims that rest on secondary citation rather than verbatim quotation.

Schmidhuber vs Hinton: what's different

The companion catalog hinton-problems emphasizes representational toy tasks: small benchmarks (4-2-4 encoder, family trees, shifter) designed to expose what kind of internal representation a network develops. Hidden-unit inspection is the experimental payoff.

Schmidhuber's lineage emphasizes algorithmic capability: long-time-lag indexing (flip-flop, chunker, adding, temporal-order, a^n b^n c^n), key-value binding (1992 fast-weights → 2021 linear Transformers), Kolmogorov-complexity search (Levin → OOPS), and controller+model+curiosity loops in tiny stochastic environments (1990 pole-balance → 2018 World Models). The signature methodological move is the controlled difficulty sweep — (q=50, p=50) → (q=1000, p=1000) in the 1997 LSTM paper, the 5,400-experiment grid in the 2017 Search Space Odyssey. See the closing thematic synthesis in docs/lineage.md if/when populated, or the per-stub What it demonstrates sections.

About

Stubs for synthetic learning problems from Jürgen Schmidhuber's papers (1989–2025) — companion to hinton-problems

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages