(This is a permanent version of this page. The latest version of this page can be found under Revision Notes, at the end.)
See experiments 2 3 4 - knowledge visualisation
These were early experiments, before I had really thought through what I was testing, and it shows - for example:
Nevertheless, I'm analysing here their results to better design experiment 4.
The single most used tag in the collection was psychology (68 resources), yet this was not mentioned as a theme or subtheme. This may be because there was too much content for ChatGPT to consume, of course.
What of the Experiment 2 mindmap: c-2-s-0-response-p2-mindmap.jpeg?
While ChatGPT identified 8 toplevel themes, the mindmap only displays 4. Not good!
Did we do better with a collection of only 10 resources?
Even with only 10 resources, it's a lot of work for me to evaluate the appropriateness of all 6 themes and 27 subthemes ChatGPT found within them (see c-3_s-0-response-p-2. At first glance the 6 themes look right, but a moment's thought and research reveals:
Moreover, when I asked it to identify subthemes which appear in more than one theme, it started making things up: the subtheme "Impact of media literacy education and fact-checking efforts" does not appear in the theme of "the need for effective communication, dialogue, and understanding in addressing misinformation and promoting accurate information", simply because that theme does not exist.
The mindmap () on the other hand, is a lot better than in Experiment 2: .
All top-level themes are present and it's easier to read. But some subthemes are missing:
Remark: 5 of these 7 subthemes were the last subtheme listed under their theme.
Also, while ChatGPT identified "Difficulty in changing beliefs and susceptibility to ideologically motivated cognition" as belonging in two themes, on the mindmap it appears to be linked to five - almost all of the 7 themes.
Using S-0 and Prompt 2, ChatGPT invented themes that were not there and ignored themes that were, and then misrepresented its (wrong) findings when visualising them.
Experiment 4 is an evolution of Experiments 2 & 3 in that the same collection and summariser was used, and one of the end goals is a visualisation of a Collection's main themes.
However, I'm aiming for a different goal, reflecting the problems discovered earlier:
So rather than trying to get an accurate visualisation of a Collection, I'm now more interested in getting a provocative one - something that will make me think by spotting potential new connections, even if those connections were not themselves present in the collection.
For this, a different mindmap is required, showing both the articles and the themes.
Hence prompt 2b - themes visualised. However:
So I turned to a new mapping tool (Graphviz) and its test neato visualisation, and a new prompt (prompt 2c - themes and resources visualised) to take advantage of it.
In c-3_allnotes-response-p-2c I started with "version 1" of Prompt2c, trying to get a useful visualisation of the same content (c-3_allnotes) using Graphviz. The result wasn't great to say the least (c-3_allnotes-response-p-2c-a-1.png):
The conversation therefore continued as I tried tweaking node size, shape and colour; introduced line breaks into the texts so they'd fit inside the nodes; and tried to make the whole thing more compact.
Near the end I made the most progress by simply telling ChatGPT what I wanted, but the struggles beforehand taught me a lot about Graphviz. It will be interesting to see whether in the future I'll simply default to "ask ChatGPT" without all that the struggle... and learning.
Eventually I had something (c-3_allnotes-response-p-2c-a-12.png) that I can actually make sense of:
So I asked my last question of the conversation: "compose a prompt combining these characteristics so that I can generate concept maps like this quicker". I then took that, incorporated everything I learnt from c-3_allnotes-response-p-2c, and created "version 2" of Prompt2c.
This I then tested using the same collection to create c-3_allnotes-response-p-2d, resulting in the c-3_allnotes-response-p-2c-b.jpeg concept map: .
Although the content analysed in the collection is identical and the prompts very similar, the resulting concept maps are very different, with far more interconnections in the concept map resulting from the earlier conversation (2c-a-12) than in (2c-b), when I tried to distil everything into a single prompt:
One obvious possible reason for the difference in connection density is the difference in themes. The two experiments show - not for the first time - that when asked to identify themes in the same content, ChatGPT will identify different themes each time.
However it is also possible that the key driver here is the length of the conversations: 2c-a-12 was the result of an 18-response conversation as I wrestled with graphviz, whereas 2c-b had only 3 responses. It doesn't necessarily make sense - the content, after all, is the same - but ChatGPT often doesn't.
Examining the graphs I noticed that 2c-a-12 has abbreviated resource titles. At a late stage of the conversation in c-3_allnotes-response-p-2c I asked it to "make the entire thing more compact", and I see now that ChatGPT turned "I read 100-and-counting articles on Fake News so you don’t have to" into "Reading 100+ Articles on Fake News".
This didn't happen in c-3_allnotes-response-p-2d, as the prompt didn't give ChatGPT much leeway, simply asking it to "Reduce node separation to make the concept map compact".
I find the figures constantly forcing me to ask questions like why is article X not connected to Theme Y, when article Z is? For example:
But these are not the best questions. They come from comparing just resource and theme titles, while the links themselves come from ChatGPT's underlying analysis of my notes rather than the full resource. Even without hallucinations, in other words, they're not that informative.
Rather than "is this correct", the question should be "what does that make me think of?"
And there is a quantitative difference here between the 2 concept maps. 2c-b is so sparse I nothing interesting jumps out to grab my attention, whereas the centrality of the Misinformation and Cognitive Biases theme in 2c-a-12 (linked to 5 resources) gives me something to work with - for example:
This is one of this wiki's pages managed with the permanent versions pattern described in Two wiki authors and a blogger walk into a bar…