We think cortex might function like a JEPA. It looks like prediction errors in layer 2/3 are not computed against input (as is the idea in predictive processing), but against a representation in latent space (i.e. like in a JEPA
arxiv.org/abs/2301.08243 or RPL
doi.org/10.1101/2025...).

Self-Supervised Learning from Images with a Joint-Embedding Predictive Architecture
This paper demonstrates an approach for learning highly semantic image representations without relying on hand-crafted data-augmentations. We introduce the Image-based Joint-Embedding Predictive Archi...