this post was submitted on 23 May 2024
73 points (92.0% liked)

Privacy

32492 readers
369 users here now

A place to discuss privacy and freedom in the digital world.

Privacy has become a very important issue in modern society, with companies and governments constantly abusing their power, more and more people are waking up to the importance of digital privacy.

In this community everyone is welcome to post links and discuss topics related to privacy.

Some Rules

Related communities

much thanks to @gary_host_laptop for the logo design :)

founded 5 years ago
MODERATORS
 

As you may know, ChatGPT collects a lot of data on the users for the improvement of their AI, but this poses risks in its own way. I was wondering whether there are privacy alternatives to ChatGPT. Perhaps on F-Droid or Aurora/PlayStore, or for Linux.

Are there any alternatives you know of? Or are there other ways to interact with ChatGPT without giving personal information, such as a privacy focussed front-end?

you are viewing a single comment's thread
view the rest of the comments
[–] anamethatisnt@lemmy.world 4 points 7 months ago (3 children)

There are private GPT solutions coming, f.e. https://www.fujitsu.com/global/products/data-transformation/data-driven/ai-test-drive/
They are aimed at companies that for compliance reasons want to self host.

I wouldn't trust an llm solution with sensitive information unless I host it myself.

[–] NoTimeLeft@lemmy.world 2 points 7 months ago (2 children)

Do you happen to know how this self-hosting would work? Can I run it at my desktop/phone or even a raspberry pi? How is the quality of generated results compared to ChatGPT?

[–] brianorca@lemmy.world 2 points 7 months ago

I can run 7B models on my laptop with its embedded GPU. Running on a phone or a Pi is possible with smaller models, but very slow. Expect good speed with a desktop Nvidea GPU. Later this year, there should be new computers with an NPU integrated to the CPU which should speed up computers that don't have a dedicated GPU. (But a GPU will still outperform them by a lot.)

70B models will run very slowly on even the best consumer hardware due to memory limitations.

[–] a4ng3l@lemmy.world 1 points 7 months ago

Typically llm are rather ressource intensive - you need beefy hardware to run those at speed. Especially if you intend to train them with your data to improve their relevance. I don’t think mobile phones or run to the mill laptops are going to be enough for any non-trivial implementations. I might be skewed by experiences on non-personal projects though.