Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse

NodeBB

  1. Home
  2. Selfhosted
  3. What can I use for an offline, selfhosted LLM client, pref with images,charts, python code execution

What can I use for an offline, selfhosted LLM client, pref with images,charts, python code execution

Scheduled Pinned Locked Moved Selfhosted
selfhosted
25 Posts 14 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • C This user is from outside of this forum
    C This user is from outside of this forum
    [email protected]
    wrote last edited by [email protected]
    #1

    I was looking back at some old lemmee posts and came across GPT4All. Didn't get much sleep last night as it's awesome, even on my old (10yo) laptop with a Compute 5.0 NVidia card.

    Still, I'm after more, I'd like to be able to get image creation and view it in the conversation, if it generates python code, to be able to run it (I'm using Debian, and have a default python env set up). Local file analysis also useful. CUDA Compute 5.0 / vulkan compatibility needed too with the option to use some of the smaller models (1-3B for example). Also a local API would be nice for my own python experiments.

    Is there anything that can tick the boxes? Even if I have to scoot across models for some of the features? I'd prefer more of a desktop client application than a docker container running in the background.

    J bjoern_tantau@swg-empire.deB andrew0@lemmy.dbzer0.comA breadsmasher@lemmy.worldB C 10 Replies Last reply
    69
    • C [email protected]

      I was looking back at some old lemmee posts and came across GPT4All. Didn't get much sleep last night as it's awesome, even on my old (10yo) laptop with a Compute 5.0 NVidia card.

      Still, I'm after more, I'd like to be able to get image creation and view it in the conversation, if it generates python code, to be able to run it (I'm using Debian, and have a default python env set up). Local file analysis also useful. CUDA Compute 5.0 / vulkan compatibility needed too with the option to use some of the smaller models (1-3B for example). Also a local API would be nice for my own python experiments.

      Is there anything that can tick the boxes? Even if I have to scoot across models for some of the features? I'd prefer more of a desktop client application than a docker container running in the background.

      J This user is from outside of this forum
      J This user is from outside of this forum
      [email protected]
      wrote last edited by
      #2

      Would like fries or a jetpack with that?

      1 Reply Last reply
      5
      • C [email protected]

        I was looking back at some old lemmee posts and came across GPT4All. Didn't get much sleep last night as it's awesome, even on my old (10yo) laptop with a Compute 5.0 NVidia card.

        Still, I'm after more, I'd like to be able to get image creation and view it in the conversation, if it generates python code, to be able to run it (I'm using Debian, and have a default python env set up). Local file analysis also useful. CUDA Compute 5.0 / vulkan compatibility needed too with the option to use some of the smaller models (1-3B for example). Also a local API would be nice for my own python experiments.

        Is there anything that can tick the boxes? Even if I have to scoot across models for some of the features? I'd prefer more of a desktop client application than a docker container running in the background.

        bjoern_tantau@swg-empire.deB This user is from outside of this forum
        bjoern_tantau@swg-empire.deB This user is from outside of this forum
        [email protected]
        wrote last edited by
        #3

        [email protected]

        1 Reply Last reply
        12
        • C [email protected]

          I was looking back at some old lemmee posts and came across GPT4All. Didn't get much sleep last night as it's awesome, even on my old (10yo) laptop with a Compute 5.0 NVidia card.

          Still, I'm after more, I'd like to be able to get image creation and view it in the conversation, if it generates python code, to be able to run it (I'm using Debian, and have a default python env set up). Local file analysis also useful. CUDA Compute 5.0 / vulkan compatibility needed too with the option to use some of the smaller models (1-3B for example). Also a local API would be nice for my own python experiments.

          Is there anything that can tick the boxes? Even if I have to scoot across models for some of the features? I'd prefer more of a desktop client application than a docker container running in the background.

          andrew0@lemmy.dbzer0.comA This user is from outside of this forum
          andrew0@lemmy.dbzer0.comA This user is from outside of this forum
          [email protected]
          wrote last edited by
          #4

          Ollama for API, which you can integrate into Open WebUI. You can also integrate image generation with ComfyUI I believe.

          It's less of a hassle to use Docker for Open WebUI, but ollama works as a regular CLI tool.

          O M C 3 Replies Last reply
          18
          • C [email protected]

            I was looking back at some old lemmee posts and came across GPT4All. Didn't get much sleep last night as it's awesome, even on my old (10yo) laptop with a Compute 5.0 NVidia card.

            Still, I'm after more, I'd like to be able to get image creation and view it in the conversation, if it generates python code, to be able to run it (I'm using Debian, and have a default python env set up). Local file analysis also useful. CUDA Compute 5.0 / vulkan compatibility needed too with the option to use some of the smaller models (1-3B for example). Also a local API would be nice for my own python experiments.

            Is there anything that can tick the boxes? Even if I have to scoot across models for some of the features? I'd prefer more of a desktop client application than a docker container running in the background.

            breadsmasher@lemmy.worldB This user is from outside of this forum
            breadsmasher@lemmy.worldB This user is from outside of this forum
            [email protected]
            wrote last edited by
            #5

            AUTOMATIC1111?

            https://github.com/AUTOMATIC1111/stable-diffusion-webui

            1 Reply Last reply
            4
            • andrew0@lemmy.dbzer0.comA [email protected]

              Ollama for API, which you can integrate into Open WebUI. You can also integrate image generation with ComfyUI I believe.

              It's less of a hassle to use Docker for Open WebUI, but ollama works as a regular CLI tool.

              O This user is from outside of this forum
              O This user is from outside of this forum
              [email protected]
              wrote last edited by [email protected]
              #6

              ChainLit is a super ez UI too. Ollama works well with Semantic Kernal (for integration with existing code) and langChain (for agent orchestration). I'm working on building MCP interaction with ComfyUI's API, it's a pain in the ass.

              1 Reply Last reply
              2
              • C [email protected]

                I was looking back at some old lemmee posts and came across GPT4All. Didn't get much sleep last night as it's awesome, even on my old (10yo) laptop with a Compute 5.0 NVidia card.

                Still, I'm after more, I'd like to be able to get image creation and view it in the conversation, if it generates python code, to be able to run it (I'm using Debian, and have a default python env set up). Local file analysis also useful. CUDA Compute 5.0 / vulkan compatibility needed too with the option to use some of the smaller models (1-3B for example). Also a local API would be nice for my own python experiments.

                Is there anything that can tick the boxes? Even if I have to scoot across models for some of the features? I'd prefer more of a desktop client application than a docker container running in the background.

                C This user is from outside of this forum
                C This user is from outside of this forum
                [email protected]
                wrote last edited by [email protected]
                #7

                I've discovered jan.ai which is far faster than GPT4All, and visually a little nicer.

                EDIT: After using it for an hour or so, it seems to crash all the time, I keep on having to reset it, and currently am facing it freezing for no reason.

                O ? 2 Replies Last reply
                10
                • C [email protected]

                  I've discovered jan.ai which is far faster than GPT4All, and visually a little nicer.

                  EDIT: After using it for an hour or so, it seems to crash all the time, I keep on having to reset it, and currently am facing it freezing for no reason.

                  O This user is from outside of this forum
                  O This user is from outside of this forum
                  [email protected]
                  wrote last edited by [email protected]
                  #8

                  I also started using this recently and it’s very plug and play. Just open and run. It’s the only client so far that feels like I could recommend to non-geeks.

                  C 1 Reply Last reply
                  1
                  • C [email protected]

                    I was looking back at some old lemmee posts and came across GPT4All. Didn't get much sleep last night as it's awesome, even on my old (10yo) laptop with a Compute 5.0 NVidia card.

                    Still, I'm after more, I'd like to be able to get image creation and view it in the conversation, if it generates python code, to be able to run it (I'm using Debian, and have a default python env set up). Local file analysis also useful. CUDA Compute 5.0 / vulkan compatibility needed too with the option to use some of the smaller models (1-3B for example). Also a local API would be nice for my own python experiments.

                    Is there anything that can tick the boxes? Even if I have to scoot across models for some of the features? I'd prefer more of a desktop client application than a docker container running in the background.

                    V This user is from outside of this forum
                    V This user is from outside of this forum
                    [email protected]
                    wrote last edited by
                    #9

                    The main limitation is the VRAM, but I doubt any model is going to be particularly fast.

                    I think phi3:mini on ollama might be an okish fit for python, since it's a small model, but was trained on python codebases.

                    C 1 Reply Last reply
                    1
                    • O [email protected]

                      I also started using this recently and it’s very plug and play. Just open and run. It’s the only client so far that feels like I could recommend to non-geeks.

                      C This user is from outside of this forum
                      C This user is from outside of this forum
                      [email protected]
                      wrote last edited by [email protected]
                      #10

                      I agree. it looks nice, explains the models fairly well, hides away the model settings nicely, and even recommends some initial models to get started that have low requirements. I like the concept of plugins but haven't found a way to e.g. run python code it creates yet and display the output in the window

                      1 Reply Last reply
                      0
                      • V [email protected]

                        The main limitation is the VRAM, but I doubt any model is going to be particularly fast.

                        I think phi3:mini on ollama might be an okish fit for python, since it's a small model, but was trained on python codebases.

                        C This user is from outside of this forum
                        C This user is from outside of this forum
                        [email protected]
                        wrote last edited by
                        #11

                        I'm getting very-near real-time on my old laptop. Maybe a delay of 1-2s whilst it creates the response

                        1 Reply Last reply
                        3
                        • C [email protected]

                          I was looking back at some old lemmee posts and came across GPT4All. Didn't get much sleep last night as it's awesome, even on my old (10yo) laptop with a Compute 5.0 NVidia card.

                          Still, I'm after more, I'd like to be able to get image creation and view it in the conversation, if it generates python code, to be able to run it (I'm using Debian, and have a default python env set up). Local file analysis also useful. CUDA Compute 5.0 / vulkan compatibility needed too with the option to use some of the smaller models (1-3B for example). Also a local API would be nice for my own python experiments.

                          Is there anything that can tick the boxes? Even if I have to scoot across models for some of the features? I'd prefer more of a desktop client application than a docker container running in the background.

                          H This user is from outside of this forum
                          H This user is from outside of this forum
                          [email protected]
                          wrote last edited by
                          #12

                          Maybe LocalAI? It doesn't do python code execution, but pretty much all of the rest.

                          C 1 Reply Last reply
                          4
                          • C [email protected]

                            I was looking back at some old lemmee posts and came across GPT4All. Didn't get much sleep last night as it's awesome, even on my old (10yo) laptop with a Compute 5.0 NVidia card.

                            Still, I'm after more, I'd like to be able to get image creation and view it in the conversation, if it generates python code, to be able to run it (I'm using Debian, and have a default python env set up). Local file analysis also useful. CUDA Compute 5.0 / vulkan compatibility needed too with the option to use some of the smaller models (1-3B for example). Also a local API would be nice for my own python experiments.

                            Is there anything that can tick the boxes? Even if I have to scoot across models for some of the features? I'd prefer more of a desktop client application than a docker container running in the background.

                            T This user is from outside of this forum
                            T This user is from outside of this forum
                            [email protected]
                            wrote last edited by
                            #13

                            You can tell Open Interpreter to run commands based on you human-language input. If you want local only LLM, you can pair it with Ollama. It works for "interactive" use where you're asked for confirmation before a command is run.

                            I set this up in a VM because I wanted a full automatic coding "agent" which can run commands without my intervention and I did not want it to blow up main system. It did not really work though because as far as I know Open Interpreter does not have a way to "pipe" a command's output back into the LLM so that it could create feedback with linters and stuff.

                            Another issue was that Starcoder2, which is the only LLM trained on permissive licensed code I could find, only has a 15B "human-like" model. The smaller models only speak code so I don't know how that would work for agentic usage and the 15B is really slow running on DDR4 CPU. I think agents are cool though so I would like to try Aider which is a supposedly good open source agent and unlike Open Interpreter is not abandonware.

                            Thanks for coming to my blabering talk, hope this might be useful for someone.

                            1 Reply Last reply
                            1
                            • C [email protected]

                              I've discovered jan.ai which is far faster than GPT4All, and visually a little nicer.

                              EDIT: After using it for an hour or so, it seems to crash all the time, I keep on having to reset it, and currently am facing it freezing for no reason.

                              ? Offline
                              ? Offline
                              Guest
                              wrote last edited by
                              #14

                              Took ages to produce answer, and only worked once on one model, then crashed since then.

                              C 1 Reply Last reply
                              0
                              • ? Guest

                                Took ages to produce answer, and only worked once on one model, then crashed since then.

                                C This user is from outside of this forum
                                C This user is from outside of this forum
                                [email protected]
                                wrote last edited by [email protected]
                                #15

                                Try the beta on the github repo, and use a smaller model!

                                1 Reply Last reply
                                0
                                • H [email protected]

                                  Maybe LocalAI? It doesn't do python code execution, but pretty much all of the rest.

                                  C This user is from outside of this forum
                                  C This user is from outside of this forum
                                  [email protected]
                                  wrote last edited by
                                  #16

                                  This looks interesting - do you have experience of it? How reliable / efficient is it?

                                  mitexleo@buddyverse.oneM 1 Reply Last reply
                                  2
                                  • C [email protected]

                                    I was looking back at some old lemmee posts and came across GPT4All. Didn't get much sleep last night as it's awesome, even on my old (10yo) laptop with a Compute 5.0 NVidia card.

                                    Still, I'm after more, I'd like to be able to get image creation and view it in the conversation, if it generates python code, to be able to run it (I'm using Debian, and have a default python env set up). Local file analysis also useful. CUDA Compute 5.0 / vulkan compatibility needed too with the option to use some of the smaller models (1-3B for example). Also a local API would be nice for my own python experiments.

                                    Is there anything that can tick the boxes? Even if I have to scoot across models for some of the features? I'd prefer more of a desktop client application than a docker container running in the background.

                                    mitexleo@buddyverse.oneM This user is from outside of this forum
                                    mitexleo@buddyverse.oneM This user is from outside of this forum
                                    [email protected]
                                    wrote last edited by
                                    #17

                                    You should try https://cherry-ai.com/ .. It's the most advanced client out there. I personally use Ollama for running the models and Mistral API for advnaced tasks.

                                    1 Reply Last reply
                                    1
                                    • C [email protected]

                                      I was looking back at some old lemmee posts and came across GPT4All. Didn't get much sleep last night as it's awesome, even on my old (10yo) laptop with a Compute 5.0 NVidia card.

                                      Still, I'm after more, I'd like to be able to get image creation and view it in the conversation, if it generates python code, to be able to run it (I'm using Debian, and have a default python env set up). Local file analysis also useful. CUDA Compute 5.0 / vulkan compatibility needed too with the option to use some of the smaller models (1-3B for example). Also a local API would be nice for my own python experiments.

                                      Is there anything that can tick the boxes? Even if I have to scoot across models for some of the features? I'd prefer more of a desktop client application than a docker container running in the background.

                                      mitexleo@buddyverse.oneM This user is from outside of this forum
                                      mitexleo@buddyverse.oneM This user is from outside of this forum
                                      [email protected]
                                      wrote last edited by
                                      #18

                                      You should try https://cherry-ai.com/ .. It's the most advanced client out there. I personally use Ollama for running the models and Mistral API for advnaced tasks.

                                      mitexleo@buddyverse.oneM C 2 Replies Last reply
                                      3
                                      • mitexleo@buddyverse.oneM [email protected]

                                        You should try https://cherry-ai.com/ .. It's the most advanced client out there. I personally use Ollama for running the models and Mistral API for advnaced tasks.

                                        mitexleo@buddyverse.oneM This user is from outside of this forum
                                        mitexleo@buddyverse.oneM This user is from outside of this forum
                                        [email protected]
                                        wrote last edited by
                                        #19

                                        It's fully open source and free (as in beer).

                                        1 Reply Last reply
                                        1
                                        • C [email protected]

                                          This looks interesting - do you have experience of it? How reliable / efficient is it?

                                          mitexleo@buddyverse.oneM This user is from outside of this forum
                                          mitexleo@buddyverse.oneM This user is from outside of this forum
                                          [email protected]
                                          wrote last edited by
                                          #20

                                          LocalAI is pretty good but resource-intensive. I ran it on a vps in the past.

                                          1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          Powered by NodeBB Contributors
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups