Reliance on internal predictive models of the world is central to many theories of human cognition. Yet it is unknown whether humans acquired multiple separate internal models, each evolved for a specific domain, or maintain a globally unified representation. Using fMRI during naturalistic experiences (movie watching and narrative listening), we show that three topographically distinct midline prefrontal cortical regions perform distinct predictive operations. The ventromedial PFC updates contextual predictions (States), the anteromedial PFC governs reference frame shifts for social predictions (Agents), and the dorsomedial PFC predicts transitions across the abstract state spaces (Actions). Prediction-error-driven neural transitions in these regions, indicative of model updates, coincided with subjective belief changes in a domain-specific manner. We find these parallel top-down predictions are unified and selectively integrated with sensory streams in the Precuneus, shaping participants' ongoing experience. Results generalized across sensory modalities and content, suggesting humans recruit abstract, modular predictive models for both vision and language. Our results highlight a key feature of human world modeling: fragmenting information into abstract domains before global integration.