Experiment 10

In this experiment I'm going to try and combine techniques developed in as many of the previous ones as possible, using the same content:

  • Collection 10: the stuff I like this year - just 5 resources, as of 2024-01-09
  • Process: I threw the best part of my kitchen sink at this one, just to see what would happen if I followed this process using both the summarised and full notes versions of the collection, with the fullnotes sent to both chatgpt3.5 and chatgpt4:
  • Full responses:
  • Analyses: Some thoughts from the following analysis went into ChatGPT integration free trial, Bullshit, Botshit and Bubbles (newsletter 3):
    • Using Summariser 0 (ie C-10-S-0-response 1):
      • the newsletter was boring, wooden and repetitive - it didn't even try, for example, to play the two articles on "bullshit" together. I should have pushed back and asked for better, but the main goal was for me to get an overview of the articles to inform the next stage (visualisations).
      • visualisations: there was an improvement from 1 to 3, but frankly the underlying analysis was insufficiently insightful for these to be of much use
      • comparing the blog posts: the second was an improvement textually, but it made up URLs, despite having them in its memory (unless they'd passed out the token window?)
    • Using my full notes through ChatGPT3.5 vs ChatGPT4:
      • GPT4's newsletter is objectively better
        • 3.5's version writes a 1-para intro before summarising each article separately, with barely any connections between them, then a conclusion.
        • GPT4's gives us a 1-para intro followed by a 5para overview and a conclusion. The overview's paragraphs link together better.
      • Both versions of GPT produced roughly similar results. My previous conclusion - that these visualisations appear more interesting/useful than they actually are - remains unchanged
      • GPT4's newsletter is objectively better
        • after (as requested) introducing the subject (open source) and summarising the Mistral announcement, it was more able to address how the 4 other articles were relevant to open source
        • on the other hand, while I asked it to "Write in a more journalistic, less academic style" it instead wrote something closer to a Mistral press release
        • moreover, it really stretched to find anything relevant to OS in one of the articles. My notes on this article do not "vividly illustrate the practical benefits of open-source AI". This is a borderline hallucination, but it's the sort of thing that can be useful if it helps you consider a connection that wasn't actually present in the original content... if you spot it.
    • Conclusion: I'll need some time to digest this, as while the GPT4 outputs were clearly better, I'm not sure they are 100x better, while GPT4 costs 100x more.