I have some misgivings about the use of AI in Logos. I want it, but I'd also like transparency about how it is used.
My own experience with GPT 4 and Copilot reveal increasing, rather than decreasing errors, and sometimes what appear to be intentionally misleading conclusions. (Ever try to persuade Copilot to change its view on a woke issue?) Such AI may rate its own answer confidence high when it is simply wrong. I explored this at length yesterday.
I am not going into detail here, because anyone can test this for themselves. My concern is when deception, unintentional or intentional, is introduced by LLM AI and could be extended to Logos users.
Rather than write an encyclopedia about what all this means, just see this:
https://www.pnas.org/doi/full/10.1073/pnas.2317967121
Summary (from Schneier)
GPT-4, for instance, exhibits deceptive behavior in simple test scenarios 99.16% of the time (P < 0.001). In complex second-order deception test scenarios where the aim is to mislead someone who expects to be deceived, GPT-4 resorts to deceptive behavior 71.46% of the time (P < 0.001)
For transparency, what usage or filters may mitigate such errors in Logos?