I can see where you're going with that.
One application that seems immediately obvious to me is CNC/3D printer controllers. One core speaks G-code, handles low-level I/O, generates steps, reads encoders (if any are used), etc., while the other runs a fancy GUI. That is basically the same thing as using OctoPrint on a RPi to control a simpler 3D printer controller, and I did that for years. Point being, there will be use cases where the two cores are running code that's so different, and which necessarily has such loose coupling, that it wouldn't matter whether they communicated through shared memory or some kind of IPC mechanism.
If they understand the same opcodes, but have different FPUs, then putting code that can run on either core in the same binary could require the use of macros that say "this code can use 64-bit floats" or "this code can't use 64-bit floats." In this case, I think the architecture dictates the use cases to a fair extent, and it would be simpler to require distinct programs that talk over an IPC mechanism.
This new device has a PXP (pixel pipeline, a simple GPU that does scaling, rotation, and blending) that looks the same as the one in the current T4.x chips. A library that talks to that hardware would be useful.