Rebuttal letter
Manuscript NEUCOM-D-25-06426
Dear Dr. Dominguez-Morales,
We sincerely thank you and the reviewers for the constructive and insightful feedback on our manuscript, Self-orthogonalizing attractor neural networks emerging from the free energy principle. We are delighted that you consider our manuscript to be of potential interest to the readership of Neurocomputing. We are encouraged by the reviewers’ overall positive evaluation of our work. We have carefully revised the manuscript and appendix to improve the biological grounding, accessibility, empirical context, and the discussion of scalability and limitations. We have also prepared additional simulations to address comments regarding scaling and computational efficiency using more realistic data.
Below, we present our point-by-point responses to each of the reviewers’ comments (reviewer comments in bold, author replies in regular text, and quoted manuscript additions in italics). Line numbers refer to the changes-tracked version of the revised manuscript (in which changes are marked in green).
We hope that the revisions and our accompanying responses make the manuscript suitable for publication in Neurocomputing. We would also be happy to respond to any further questions or comments that you or the reviewers may have.
Best regards,
Tamas Spisak and Karl Friston
Point-by-point responses¶
Reviewer #2¶
The manuscript presents a novel theoretical framework that derives self-organizing attractor neural networks from the free energy principle, offering a biologically plausible and mathematically rigorous approach to understanding emergent inference and learning dynamics. The following issues are suggested to be considered.
Response
We thank the reviewer for acknowledging the novelty, biological plausibility, and mathematical rigor of our approach and for the insightful suggestions that we address below.
Comment 1¶
The framework claims biological plausibility, further discussion is needed on how the derived learning rules (e.g. Hebbian/anti-Hebbian) map to known neurobiological mechanisms.
Response
We thank the reviewer for this important suggestion. We now provide a more detailed discussion of the neurobiological correspondence of the derived plasticity rule, quoted below for your convenience:
New discussion paragraph (lines 543-561):
The neurobiological mechanisms of synaptic plasticity are best understood and experimentally most thoroughly studied at the level of single neurons, although efforts exist to scale up these findings to the level of populations of neurons (e.g., Robinson (2011)Huang & Lin (2023)). We believe that our framework holds promise for a better understanding of plasticity mechanisms independent of scale, as it mathematically survives arbitrary coarse-graining under the deep particular partition formalism. At the single-neuron limit, the rule reduces to a discrete-time binary Hebbian/anti-Hebbian update (formally recovered when precision → ∞), closely resembling spike-timing–dependent plasticity (STDP) Roberts & Leen, 2010Vignoud et al., 2024, where correlated activity produces long-term potentiation (LTP ~ Hebbian term) and predicted activity leads to long-term depression (LTD ~ anti-Hebbian term; see e.g. Caporale & Dan (2008)). Thereby, our framework connects STDP to predictive coding Saponati & Vinck, 2023, in the sense that presynaptic activity that is reliably predicted by postsynaptic firing is eventually depressed. Moreover, the subtractive predictive error term, together with the bounded nature of the continuous Bernoulli distribution, prevent runaway potentiation, functioning analogously to homeostatic and metaplasticity mechanisms Turrigiano, 2011Zenke et al., 2017. In sum, rather than fitting neuron-level data post hoc, our framework predicts that Hebbian/anti-Hebbian-style updates should appear at all descriptive scales, with differences only in implementation, not mathematical structure. The biological literature at synaptic, circuit, and network levels is largely consistent with this multiscale interpretation.
Comment 2¶
The simulations focus on small-scale networks. Scaling to larger network or real-world datasets would strengthen pratical relevance.
Response
We fully agree that scaling evidence strengthens practical relevance. The original reference implementation prioritized transparency of the local update rules over speed; we retain it for all core analyses (Simulations 1–4). To address scalability, we now provide three new contributions: (i) a parallelized JAX implementation for runtime profiling, (ii) systematic scaling experiments (up to 2.5B parameters), and (iii) a new simulation on a more realistic dataset. We also added a discussion paragraph that identifies the sequential update bottleneck and outlines paths to further scaling (dedicated hardware and weight sparsity). The analytical complexity comparison (per-step cost, memory capacity bounds) is detailed in response to Reviewer #3 Comment 2 (Appendix 8); here we focus on the empirical evidence.
New methods paragraph (lines 488–494):
In addition to the reference Python implementation used for Simulations 1–4, we provide a vectorized JAX implementation that applies the same inference and learning rules (eq. (14) (17)) in a parallelized, full-network update schedule (Simulation Notebook 5). This parallel schedule is computationally more efficient but not strictly equivalent to the sequential node-by-node updates of the reference implementation; we validate that both implementations produce qualitatively similar coupling matrices and retention behavior on shared test cases. The JAX implementation is used exclusively for runtime profiling (Simulation 5) and for the larger-scale face recognition experiment (Simulation 6).
New Simulation 5 paragraph (lines 487–503):
Simulation 5: scalability profile and memory capacity
We profile the computational scaling of the JAX implementation across network sizes ranging from to nodes, i.e up to 2.5B total parameters (Simulation Notebook 5). Training and inference runtimes scale as expected from the analytical per-step complexity derived in Appendix 8 (Figure 7A). Throughput remains high for moderate network sizes and drops off as expected for very large (Figure 7B). We further probe memory capacity by varying the number of stored patterns at fixed network size, measuring both retention (deterministic attractor recovery, []Figure 7C) and noisy reconstruction quality (Figure 7D). Consistent with the analytical prediction that emergent orthogonalization approaches the projector-network capacity limit (Appendix 8), we observe that the network maintains high-fidelity attractors and effective Bayesian retrieval for pattern counts substantially exceeding the classical Hopfield bound of .
New Simulation 6 paragraph (lines 504–516):
Simulation 6: face recognition with the Olivetti faces dataset
To move beyond the small-scale handwritten digit experiments, we test the framework on the Olivetti faces dataset — a benchmark consisting of 400 grayscale face images (40 subjects, 10 images each) at full resolution pixels ( nodes, more than 16 million parameters), using the full set of 400 patterns (Simulation Notebook 6). Training confirms that the key phenomena observed on handwritten digits — emergent orthogonalization and attractor formation — transfer to this substantially higher-dimensional and more naturalistic stimulus domain (Figure 7E). We further evaluate stochastic reconstruction from heavily corrupted input. As shown on Figure 7F, the network performs Bayesian inference by integrating the noisy sensory evidence (likelihood) with its learned prior beliefs (attractors), with the balance controlled by the precision (inverse temperature) parameter. Notably, the input noise level used here is severe enough that the corrupted faces approach the limit of human recognizability, yet the network reliably recovers the identity of the original face. The combined scaling, capacity, and face recognition results are shown in Figure 7.

Figure 7: Scaling, memory capacity, and face recognition. A–B: Runtime (A) and throughput (B) of the JAX implementation as a function of network size , confirming the expected per-step scaling (Simulation Notebook 5). C: Deterministic retrieval quality (retention correlation, blue) as a function of the number of stored patterns , compared with the maximum inter-pattern self-correlation baseline (red). The network maintains high-fidelity attractors well beyond the classical Hopfield capacity bound. D: Stochastic (Bayesian) reconstruction: correlation of the reconstructed output with the original pattern (blue) versus correlation of the noisy input with the original (red), as a function of . The network consistently improves upon the noisy input, demonstrating effective Bayesian retrieval. E: Training the network on the Olivetti faces dataset (, , 400 patterns; Simulation Notebook 6). Random examples of training faces (left) and the corresponding learned attractors (right), confirming attractor formation and orthogonalization in this naturalistic domain. F: Bayesian face reconstruction from noisy input. Columns show reconstructions at increasing likelihood precision (left to right, top row) and increasing prior precision (top to bottom). Even when the input is degraded to a level approaching the limit of human face recognition, the network recovers the original identity by combining sensory evidence with its learned attractor-based priors.
New discussion paragraph on scaling prospects (lines 704–714):
Finally, a key computational bottleneck is the sequential dependency of the Gibbs-like node-by-node inference and learning: each node’s new state depends on the current states of all others, preventing straightforward parallelization. Our synchronous (full-network) JAX implementation sidesteps this by applying all updates simultaneously, yielding favorable empirical scaling up to (Simulation 5); however, this parallel schedule is an approximation whose fidelity at very large warrants further study. To efficiently scale the framework towards billions of neurons two complementary directions appear most promising: (i) dedicated hardware — in particular thermodynamic computers Melanson et al., 2025 and memristive substrates Lin et al., 2023Lin et al., 2024 — which can natively implement the local, stochastic update rules without the sequential bottleneck; and (ii) structured weight sparsity, which would reduce the per-step cost from to (where is the average number of non-zero connections per node) and simultaneously lower the memory footprint.
Comment 3¶
The manuscript is well-written but occasionally dense. It is suggested to add a high-level schematic of the framework and simplify derivations where possible.
Response
We appreciate this feedback and agree. We made multiple changes to improve accessibility. First, we added an intuition-first high-level pipeline schematic (Figure 1, see below for convenience). Second, we substantially simplified the derivation of the inference and learning rules. The inference section (form line 191) now uses a compact local VFE argument — three equations from the local evidence to the inference rule — rather than the longer accuracy–complexity decomposition, which is retained in the appendix (Appendix 6) for interested readers. The learning rule derivation (from line 222) is similarly streamlined via the same local VFE. Third, we added a new intuition-focused subsection (lines 336-379) that explains the emergence of solenoidal flows from asymmetric couplings — using a dissipative-plus-rotational decomposition — without requiring a formal proof of the divergence-free condition. We revised the manuscript at many other locations to improve the narrative, see our response to Reviewer #3 Comment 5 for details.

Figure 1: Overview of the framework. Starting from universal and parsimonious structural assumptions (deep particular partition), variational free energy minimization under the Free Energy Principle (FEP) gives rise to emergent dynamics at multiple scales: local stochastic inference and learning rules that guide the update of internal states and the couplings between them, self-organizing attractor dynamics at the macroscale, and — at the computational level — Bayesian inference, approximately orthogonal attractor representations of latent external causes, and sequence learning capabilities.
Reviewer #3¶
This paper presents a highly novel and conceptually ambitious framework that derives self-organizing attractor neural networks directly from the free energy principle, without imposing ad hoc learning or inference rules. The unification of attractor dynamics, Bayesian active inference, and learning as emergent properties of a single variational objective is a significant theoretical contribution. The result that attractors self-orthogonalize to balance predictive accuracy and model complexity is particularly original and provides a principled explanation for efficient representation, generalization, and information maximization. The distinction between symmetric couplings under random inputs and asymmetric, non-equilibrium dynamics under sequential inputs is insightful and offers a compelling theoretical bridge to, and extension of, Boltzmann Machines.
Response
We thank the reviewer for highlighting the significance and originality of our work and for the insightful comments that we address below point by point.
Comment 1¶
Explicitly discuss how orthogonalized attractor representations and free-energy minimization affect robustness to adversarial perturbations, noise, or data poisoning—topics of growing importance in secure AI.
Response
We thank the reviewer for raising this stimulating point — thinking about the framework through the lens of adversarial robustness is a genuinely interesting perspective. We now address it in a new discussion subsection and a supporting mathematical appendix (Appendix 7).
New discussion subsection (lines 642–658):
A natural question is whether the structural properties of FEP-based self-orthogonalizing attractor networks confer robustness to adversarial perturbations, noise corruption, or data poisoning — topics of growing importance in secure and trustworthy AI Goodfellow et al., 2014. Within the Bayesian framing of the network (eq. (20)), adversarial perturbations and data poisoning correspond to attacks at two distinct levels: sensory perturbations (including adversarial examples) enter as erroneous bias shifts during inference (corrupted likelihood), whereas data poisoning distorts the learned prior , i.e. the attractor landscape itself. The inverse-temperature parameter mediates the balance between these levels: high precision deepens attractor basins relative to any bias perturbation, so the magnitude of becomes small compared to the basin depth; low precision, conversely, flattens the prior landscape and lets the actual sensory evidence dominate, mitigating the effect of distorted or poisoned attractors. This precision-mediated trade-off is a direct consequence of the Bayesian posterior structure and has no counterpart in deterministic Hopfield-type retrieval (see Appendix 7 for formal analysis). Furthermore, our framework naturally distinguishes between two types of redundancy. Self-orthogonalization minimizes representational redundancy (attractor overlap) which maximizes inter-attractor distances and thereby the “adversarial budget”, i.e. the minimum perturbation energy needed to push the network across a basin boundary (Appendix 7). On the other hand, the network maintains high structural redundancy: the attractors are supported by distributed weights, so bounded weight corruption or node failure has limited impact on the energy landscape.
The formal analysis is provided in a new appendix (Appendix 7), which derives how attractor orthogonality maximizes the adversarial budget (minimum perturbation norm to cross a basin boundary), how the precision parameter mediates the trade-off between resilience to sensory attacks and data poisoning, and how stochastic MCMC averaging and the anti-Hebbian learning term provide additional implicit regularization.
Comment 2¶
Include clearer empirical or analytical comparisons with established models (e.g., Hopfield networks, Boltzmann Machines, predictive coding models) to better contextualize the practical gains.
Response
We agree that a structured comparison strengthens the contribution. We added a short discussion paragraph pointing to a detailed analytical appendix (Appendix 8) that includes a comparison table and formal complexity and mixing-time analyses (see below for convenience).
New discussion paragraph (lines 622–632):
Relative to classical single-layer Hopfield/Boltzmann formulations, the present framework preserves attractor-based inference while extending the model class along several dimensions that emerge from the FEP derivation rather than being imposed by design. These extensions — and their analytical consequences — are elaborated in Appendix 8. Two points deserve emphasis here. First, the learning rule (eq. (17)) eliminates the costly free-running phase of contrastive divergence, reducing per-step learning complexity from to (Appendix 8). Second, in the presence of asymmetric couplings the solenoidal component of the dynamics induces non-reversible probability currents that accelerate mixing relative to symmetric Boltzmann machines — a mechanism formally analogous to continuous normalizing flows Tong et al., 2023 and known to reduce mixing times Ao, 2004Xing, 2010 (Appendix 8). The relationship to hierarchical predictive coding — which minimizes the same objective in a bidirectional hierarchy rather than a fully recurrent topology — is discussed in Appendix 8 and constitutes a promising direction for future work.
The detailed analytical comparison is provided in a new appendix (Appendix 8), covering learning-rule complexity (elimination of the free-running phase, reducing per-step cost from to ), solenoidal flows and mixing-time acceleration, progressive orthogonalization and memory capacity approaching the projector-network limit, and the relationship to hierarchical predictive coding. The summary comparison table is reproduced below for convenience:
Table 1:Comparison of the FEP-ANN with canonical formulations of classical models.
Feature Hopfield Boltzmann machine Projector network Predictive coding FEP-ANN (ours) Derivation Energy-based Statistical mechanics Optimal storage Hierarchical Bayesian First principles (FEP) State space Binary Binary Binary Continuous (Gaussian) Continuous (CB) Activation Sign Logistic sigmoid Sign Linear / nonlinear Langevin (emergent) Learning Hebbian (batch) Contrastive divergence Pseudo-inverse (batch) Prediction-error min. Hebbian/anti-Hebbian (online) Learning phases One-shot Clamped + free One-shot Layerwise Single phase Coupling Symmetric Symmetric Symmetric Directed (hierarchy) Symmetric or asymmetric Sequence dynamics No No No Temporal hierarchy NESS solenoidal flow Orthogonalization No No By construction No Emergent (VFE) Memory capacity (optimal) N/A (generative) Approaches Inference Deterministic Gibbs MCMC Deterministic Message passing MCMC + solenoidal flow Precision control None Temperature (fixed) None Precision-weighting (Bayesian) Continual learning Catastrophic forgetting Catastrophic forgetting Catastrophic forgetting Possible (with replay) Built-in (spontaneous replay) Learning complexity one-shot per step one-shot per layer per step per step FEP / active inference No formal link Variational (post hoc) No formal link Compatible Derived from FEP
Comment 3¶
Address computational scalability and potential limitations when applied to high-dimensional or real-world data streams.
Response
We now address scalability both analytically and empirically. The analytical treatment — per-step complexity, memory capacity bounds, and mixing-time considerations — is provided in Appendix 8 (see also our response to Reviewer #3 Comment 2). The empirical scaling evidence — runtime profiling across network sizes, memory capacity sweeps, and a new face recognition experiment — is detailed in our response to Reviewer #2 Comment 2. We also added a new discussion paragraph (quoted in Reviewer #2 Comment 2) that identifies the sequential update dependency as the principal bottleneck and outlines two concrete paths to scale.
Revised discussion paragraph (lines 679–687):
The per-step computational cost of the FEP-based attractor network is for a full-network update — dominated by one matrix-vector product for inference and one outer-product update for learning — with no free-running phase required (unlike contrastive divergence in Boltzmann machines; see Appendix 8). This complexity is confirmed empirically: runtime scales quadratically with across network sizes from 64 to 4096 nodes (Simulation Notebook 5, Appendix 9). Memory capacity benefits from emergent orthogonalization, which progressively approaches the projector-network limit of , substantially exceeding the classical Hopfield bound of (Simulation Notebook 5). We further validated the framework on the Olivetti faces dataset at full resolution (, 400 patterns; Simulation Notebook 6), confirming that orthogonalization, Bayesian retrieval, and generalization transfer to more naturalistic stimuli.
Revised discussion paragraph (lines 694–704):
Despite the promise, several limitations in terms of scalability warrant explicit acknowledgment First, the memory footprint for the full weight matrix becomes prohibitive for very large ; sparse or factored weight parametrizations are a natural extension but remain unexplored. Second, while solenoidal flows from asymmetric couplings are expected to accelerate mixing (Appendix 8), a rigorous characterization of mixing times as a function of network size, pattern count, and coupling asymmetry is needed. Third, the current demonstrations use stationary or slowly varying input statistics; performance under strongly non-stationary, streaming real-world data — where the input distribution shifts faster than the learning rate can track — remains an open question. Fourth, all simulations use a single-layer architecture; scaling to hierarchical (multi-layer) deep particular partitions, which could support richer generative models, is an important theoretical and practical frontier.
A new appendix (Appendix 9) reports runtime benchmarks (see ) and memory capacity results for the JAX implementation across network sizes from to . The combined scaling, capacity, and face recognition results are presented in Figure 7.
Comment 4¶
When the author talks about neural networks, these new types of neural network models should be discussed, for example a multiwing hyperchaotic memristive neural network, Memristive CNN with multi-butterfly attractors, and so on.
Response
We thank the reviewer for this suggestion. We have now extended the existing discussion paragraph on thermodynamic and neuromorphic computing to include memristive and hyperchaotic attractor-network families. These systems demonstrate that recurrent, state-dependent couplings can generate rich attractor repertoires (multistability, hyperchaos, coexisting attractors) in hardware substrates whose physics naturally supplies the stochastic fluctuations that our framework requires. Our contribution is orthogonal in emphasis: we derive the architecture from first principles (FEP and deep particular partitions), rather than prescribing a phenomenological attractor family a priori.
Below we quote the revised passages:
Revised paragraph (lines 659–669):
Stochasticity - a key property of our network - is also very relevant from the perspective of artificial intelligence research. In our framework, noise is not an enemy; it implements the precision of inference, allowing it to strike a balance between stability and flexibility. This inherent stochasticity yields an exceptional fit with energy-efficient neuromorphic architectures Schuman et al., 2022, particularly within the emerging field of thermodynamic computing Melanson et al., 2025 and memristive technologies, where attractor networks — including memristive Hopfield networks, multi-wing and multi-butterfly hyperchaotic constructions, and memristive cellular-neural-network-like circuits Lin et al., 2023Wang et al., 2023Lin et al., 2024Deng et al., 2025Diao et al., 2024 have recently been studied. A promising direction is to ask whether local VFE-minimizing updates can be instantiated in thermodynamic and memristive substrates, combining the principled inferential interpretation of the FEP framework with the hardware efficiency and rich non-equilibrium dynamics in these emerging paradigms.
Revised passage (lines 709–714):
To efficiently scale the framework towards billions of neurons two complementary directions appear most promising: (i) dedicated hardware — in particular thermodynamic computers Melanson et al., 2025 and memristive substrates Lin et al., 2023Lin et al., 2024 — which can natively implement the local, stochastic update rules without the sequential bottleneck; and (ii) structured weight sparsity, which would reduce the per-step cost from to (where is the average number of non-zero connections per node) and simultaneously lower the memory footprint.
Comment 5¶
While mathematically elegant, some concepts (e.g., “universal partitioning of random dynamical systems”) would benefit from additional intuitive explanation to broaden accessibility.
Response
We appreciate this recommendation and agree. We made several targeted changes to improve intuitive accessibility without sacrificing mathematical precision:
High-level schematic Figure 1: A new pipeline figure shows the logical flow from structural assumptions (deep particular partition) through local FEP dynamics to emergent computational properties (see also Reviewer #2 Comment 3).
Universal partitioning of random dynamical systems: We added intuition sentences at multiple points (e.g., lines
109, 121, 166, 202, 204, 248, 328, 353, 371). For instance, we clarify that universal partitioning means that any persistent random dynamical system can be read as performing inference at its boundaries — i.e., nested interfaces separate what is inferred from what is inferred about, so the system has a natural interpretation as an inference engine without ad hoc modeling devices.Simplified derivations (lines
182-189, 193–210, 221-226): We restructured the main results to emphasize the local evidence form , from which both the inference rule and the learning rule follow in a few lines via local VFE minimization. The longer accuracy–complexity decomposition is now in Appendix 6. Crucially, the local derivation no longer requires the global Boltzmann joint as an intermediate step — it proceeds directly from the deep particular partition structure to the update rules.Self-solenoidization intuition (, lines
336–379): We added a new subsection that explains how asymmetric couplings induce sequence learning via solenoidal flows using an intuitive dissipative-plus-rotational decomposition. The key insight — that attractor orthogonality drives probability currents along iso-energy contours, is presented heuristically, making the concept accessible without full technical detail.Reframed global joint (line
368): The Boltzmann-like stationary distribution is now introduced as a consequence of the local pairwise structure (“As a global consequence of the local pairwise structure...”), making clear that it is a derived property rather than an assumption.
We believe these changes substantially lower the entry barrier while preserving the full mathematical content in the appendix for readers who want it.
- Robinson, P. A. (2011). Neural field theory of synaptic plasticity. Journal of Theoretical Biology, 285(1), 156–163. 10.1016/j.jtbi.2011.06.023
- Huang, C.-H., & Lin, C.-C. K. (2023). New biophysical rate-based modeling of long-term plasticity in mean-field neuronal population models. Computers in Biology and Medicine, 163, 107213. 10.1016/j.compbiomed.2023.107213
- Roberts, P. D., & Leen, T. K. (2010). Anti-Hebbian Spike-Timing-Dependent Plasticity and Adaptive Sensory Processing. Frontiers in Computational Neuroscience, 4. 10.3389/fncom.2010.00156
- Vignoud, G., Venance, L., & Touboul, J. D. (2024). Anti-Hebbian plasticity drives sequence learning in striatum. Communications Biology, 7(1). 10.1038/s42003-024-06203-8
- Caporale, N., & Dan, Y. (2008). Spike Timing–Dependent Plasticity: A Hebbian Learning Rule. Annual Review of Neuroscience, 31(1), 25–46. 10.1146/annurev.neuro.31.060407.125639
- Saponati, M., & Vinck, M. (2023). Sequence anticipation and spike-timing-dependent plasticity emerge from a predictive learning rule. Nature Communications, 14(1). 10.1038/s41467-023-40651-w
- Turrigiano, G. (2011). Homeostatic Synaptic Plasticity: Local and Global Mechanisms for Stabilizing Neuronal Function. Cold Spring Harbor Perspectives in Biology, 4(1), a005736–a005736. 10.1101/cshperspect.a005736
- Zenke, F., Gerstner, W., & Ganguli, S. (2017). The temporal paradox of Hebbian learning and homeostatic plasticity. Current Opinion in Neurobiology, 43, 166–176. 10.1016/j.conb.2017.03.015
- Melanson, D., Abu Khater, M., Aifer, M., Donatella, K., Hunter Gordon, M., Ahle, T., Crooks, G., Martinez, A. J., Sbahi, F., & Coles, P. J. (2025). Thermodynamic computing system for AI applications. Nature Communications, 16(1). 10.1038/s41467-025-59011-x
- Lin, H., Wang, C., Yu, F., Sun, J., Du, S., Deng, Z., & Deng, Q. (2023). A Review of Chaotic Systems Based on Memristive Hopfield Neural Networks. Mathematics, 11(6), 1369. 10.3390/math11061369
- Lin, H., Deng, X., Yu, F., & Sun, Y. (2024). Grid Multibutterfly Memristive Neural Network With Three Memristive Systems: Modeling, Dynamic Analysis, and Application in Police IoT. IEEE Internet of Things Journal, 11(18), 29878–29889. 10.1109/jiot.2024.3409373
- Goodfellow, I. J., Shlens, J., & Szegedy, C. (2014). Explaining and Harnessing Adversarial Examples. arXiv. 10.48550/ARXIV.1412.6572
- Tong, A., Fatras, K., Malkin, N., Huguet, G., Zhang, Y., Rector-Brooks, J., Wolf, G., & Bengio, Y. (2023). Improving and generalizing flow-based generative models with minibatch optimal transport. arXiv. 10.48550/ARXIV.2302.00482
- Ao, P. (2004). Potential in stochastic differential equations: novel construction. Journal of Physics A: Mathematical and General, 37(3), L25–L30. 10.1088/0305-4470/37/3/l01
- Xing, J. (2010). Mapping between dissipative and Hamiltonian systems. Journal of Physics A: Mathematical and Theoretical, 43(37), 375003. 10.1088/1751-8113/43/37/375003