I am finding myself turning to gpt-4o-mini a whole lot more since they added prompt caching last month - where you get an automated 50% discount if you send the same tokens twice or more
It is fantastic for use-cases like answering questions about a medium sized codebase
@simon Curious, how do you feed the whole codebase into the context?