• Subscribe
  • would you prefer an local llm software?

    Roche
    1 reply
    there are many online llm ui, like monica,maxai would you prefer local llm chat with pdf/chat with something you like? if yes, which model or ability you want?

    Replies

    Ruben Boonzaaijer
    I use ollama to host my local models like mistral and llama2
    Share