I am using Mistral 7b Instruct for text summary and some light "assistant" type chatting for the last several months. I have been pleased at how accurate it is for my needs, especially given it's size.
I recall alot of trial and error to find models that were compatible with the version of llama-cpp-python that oobabooga uses (at any given time). GGUF should have made the model format (and therefore model selection) more simple, but i imagine there are still nuances that make it more difficult than it should be to find a working model for a noob.