Had a thought, and I must put it out here: Over the past year, building with LLMs, I ve noticed that the big change is not necessarily that AI gives us better answers. Change your perspective and think like this: we, humans, are starting to ask better questions. And here's why: When search engines dominated, we asked: What is X? How do I do Y? Best tools for Z? With LLMs, the questions are different: Help me think through this decision. Challenge my assumptions. What am I not seeing? Act as a product strategist and critique this. Clarify my messy thinking. That s not getting the information, that s cognitive augmentation. I feel we re moving from: Finding knowledge to extending reasoning And this raises bigger questions for builders: 1. Are we designing AI tools that optimize for answers or for better thinking? 2. Is the real advantage having a stronger model... or asking better questions? 3. In a world where everyone has access to GPT-level intelligence, does leverage come from prompting skill, system design, or taste? 4. What new human skills become valuable when execution becomes cheap? It feels like we re entering a phase where: Curiosity > credentials Clarity > speed Systems thinking > single outputs If you re building in AI right now: What types of questions are your users asking that they couldn t (or wouldn t) ask before? And are you building for answers or for thinking? Would love to hear what patterns you're seeing. ______________________ www.zackapp.space Download Zack from the App Store and soon from Google Play: https://apps.apple.com/us/app/za...
The yolog v0.5.1 build has been uploaded. All Pro features are now free. I will update the website soon. A more detailed release announcement will come later.