The chatbot gave wildly different answers to the same math problem, with one version of ChatGPT even refusing to show how it came to its conclusion.
Can we discuss how it's possible that the paid model (gpt4) got worse and the free one (gpt3.5) got better? Is it because the free one is being trained on a larger pool of users or what?
Today I used Bing Chat to get some simple batch code.
The two answers I got were wrong.
But in each response the reference link had the correct answer. ¯_(ツ)_/¯