There's really no good interface for humans to perceive knowledge in multi-dimensional, non-linear relationships, on 2D screens and 1D scrolls.
Today, websites allow users to randomly drop themselves on their knowledge graph and then perceive everything within a half-mile radius. But it’s still not possible to sense the global structure.
When interfaces typically project graph-like knowledge on the screen, they usually default to flattening these complex, multi-dimensional structures into lists and tables. However, in doing so, they lose the "relational" essence that makes a graph valuable. In getting familiar with the app, users need to identify implicit relations within the application themselves.
Graph-like knowledge is typically accessed by searching the node attributes. Usually, search interfaces have strong limitations on the attributes allowed in the search. Sometimes, it’s funny and indicative of this, when Google Search outperforms Twitter and Reddit on their own content.
Displaying more than one node side by side for comparison has been useful in code review and online retail. They reveal relationships, one at a time. Switching one node with another of a certain choice still relies upon the user guiding the navigation on a path they do not see.
Showing breadth-first or depth-first relationships on a side panel helps as an interface for making exploration and exploitation tradeoffs, but it doesn’t solve the problem. Cases like showing recommendations and similar products on a side panel.
At the block level, complex structures are projected by bullet point lists and tables. When the number of nodes is significantly larger than the number of relations, tables and nested lists work really great. But they fail as the number of relations grows from one to three and beyond.
Chat panels have opened a new option to pick the most relevant nodes from the tree, regardless of whether they are on the user’s current path. They are making surprises happen these days, but at the expense of imposing significantly more effort from the user in typing.
In ChatGPT-style interfaces, I'm not sure the barrier of implicit knowledge has been lowered as much as it has been displaced with new ones. (eg, now people have to figure out what prompts work, run it by a council, etc.)