Preferences

trilogic parent (dead)
[flagged]

woadwarrior01
> If you looking for privacy there is only 1 app in the whole wide internet right now, HugstonOne

That's a tall claim.

I've been selling a macOS and iOS private LLM app on the App Store for over two years now, that is:

a) is fully native (not electron.js) b) not a llama.cpp / MLX wrapper c) fully sandboxed (none of Jan, Ollama, LM Studio are)

I will not promote. Quite shameless of you to shill your electron.js based llama.cpp wrapper here.

threecheese
Purchased, to show my support (and to play around ofc).
woadwarrior01
Thanks! Also check out my other free privacy focussed app. :)
trilogic OP (dead)
imiric
You mean this[1]?

It's not open source, has no license, runs on Windows only, and requires an activation code to use.

Also, the privacy policy on their website is missing[2].

Anyone remotely concerned about privacy wouldn't come near this thing.

Ah, you're the author, no wonder you're shilling for it.

[1]: https://github.com/Mainframework/HugstonOne

[2]: https://hugston.com/privacy

trilogic OP
The app has a licence well visible when you install the app. The rest is written in the website and in github. Then about: requires an activation code to use, ofc it is made for ethical research purposes, so yes I am distributing it responsibly. And you can see the videos in the youtube channel for how it works. But the most important point is that you can try it easily with a firewall to see that it do not leak bytes like all the rest there. That´s what i call privacy, It has a button that cut all connection. You can say what you want but that´s it that´s all.
do_not_redeem
> But the most important point is that you can try it easily with a firewall to see that it do not leak bytes like all the rest there.

Great to hear! Since you care so much about privacy, how can I get an activation code without sending any bytes over a network or revealing my email address?

riquito
Closed source, without 3rd party independent review and people should just trust you? As if your app cannot start sending data away in a month or attempt to detect monitoring software, to name a couple
kgeist
>I challenge everyone to find another local GUI with that privacy

Llama.cpp's built-in web UI.

trilogic OP
This is from webui website docs: Once saved, Open WebUI will begin using your local Llama.cpp server as a backend! So you see Llama server not CLI. That´s a big flag there. I repeat no app in the whole world takes seriously privacy like HugstonOne. This is not advertisement, I am just making a point.
kgeist
I'm not sure what you're talking about. Llama.cpp is an inference server which runs LLMs locally. It has a built-in web UI. You can't get more private than the inference server itself.

I tried downloading your app, and it's a whopping 500 MB. What takes up the most disk space? The llama-server binary with the built-in web UI is like a couple MBs.

trilogic OP
With all respect you do seem to not understand much of how privacy works. Llama-server is working in Http. And yes the app is a bit heavy as is loading llm models using llama.cpp cli and multimodal which in itself are quite heavy, also just the dlls for cpu/gpu are huge, (just the one for the nvidial gpu is 500mb if I don't go wrong).
kgeist
Unless you expose random ports on the local machine to the Internet, running apps on localhost is pretty safe. Llama-server's UI stores conversations in the browser's localStorage so it's not retrievable even if you expose your port. To me, downloading 500 MB from some random site feels far less safe :)

>the app is a bit heavy as is loading llm models using llama.cpp cli

So it adds an unnecessary overhead of reloading all the weights to VRAM on each message? On some larger models it can take up to a minute. Or you somehow stream input/output from an attached CLI process without restarting it?

rcakebread
Says the guy with a link to a broken privacy policy on their website.
trilogic OP
I accept critics, and I thank you for it. It will be fixed ASAP.
giantrobot
> With all respect you do seem to not understand much of how privacy works. Llama-server is working in Http.

What in the world are you trying to say here? llama.cpp can run completely locally and web access can be limited to localhost only. That's entirely private and offline (after downloading a model). I can't tell if you're spreading FUD about llama.cpp or are just generally misinformed about how it works. You certainly have some motivated reasoning trying to promote your app which makes your replies seem very disingenuous.

This item has no comments currently.