What is a self-hosted small LLM actually good for (<= 3B)
-
I've tried coding and every one I've tried fails unless really, really basic small functions like what you learn as a newbie compared to say 4o mini that can spit out more sensible stuff that works.
I've tried explanations and they just regurgitate sentences that can be irrelevant, wrong, or get stuck in a loop.
So. what can I actually use a small LLM for? Which ones? I ask because I have an old laptop and the GPU can't really handle anything above 4B in a timely manner. 8B is about 1 t/s!
Learning/practice, and any use that feeds in sensitive data you want to keep on-prem.
Unless you’re set to retire within the next 5 years, the best reason is to keep your resume up to date with some hands-on experience. With the way they’re trying to shove AI into every possible application, there will be few (if any) industries untouched. If you don’t start now, you’re going to be playing catch up in a few years.
-
Please be very careful. The python code it'll spit out will most likely be outdated, not work as well as it should (the code isn't "thought out" as if a human did it.
If you want to learn, dive it, set yourself tasks, get stuck, and f around.
I know what you mean. All the code generated with ai was loaded with problems. Specifically it kept forcing my api keys into the code without using environmental variables. But for basic coding concepts it has so far been perfect. even a 3b model seemingly generates great definitions
-
So this wasn't a post actually asking what a small LLM was good for, it was just an opportunity you could use to dump on LLM usage I take it. So this whole thing was made in bad faith?
With the comments about "vibe coding" and such, all it looks like you're doing here is arguing the "merits" of how it's being used, and you're not interested in its actual usage at all.
Nobody is being pissy here except you. Small LLMs can be used for tasks such as this, and it doesn't have to be twitch - It could be an assistant that you build for reminders in your personal life - using it on twitch is a minor detail that you seem to have latched onto because you just want to dump on LLM usage.
Go to /c/fuck_ai for that.
I gave you an example that it's good for, and all you want to do is argue the merits of how I'm using it (even though it falls perfectly within Twitches TOS and use cases)
You're conflating me asking how to use these tools with you who's misusing them. I see you still don't accept what you're doing is wrong. But go you.
-
I've tried coding and every one I've tried fails unless really, really basic small functions like what you learn as a newbie compared to say 4o mini that can spit out more sensible stuff that works.
I've tried explanations and they just regurgitate sentences that can be irrelevant, wrong, or get stuck in a loop.
So. what can I actually use a small LLM for? Which ones? I ask because I have an old laptop and the GPU can't really handle anything above 4B in a timely manner. 8B is about 1 t/s!
I have it roleplay scenarios with me and sometimes I verbally abuse it for fun.
-
I've tried coding and every one I've tried fails unless really, really basic small functions like what you learn as a newbie compared to say 4o mini that can spit out more sensible stuff that works.
I've tried explanations and they just regurgitate sentences that can be irrelevant, wrong, or get stuck in a loop.
So. what can I actually use a small LLM for? Which ones? I ask because I have an old laptop and the GPU can't really handle anything above 4B in a timely manner. 8B is about 1 t/s!
for coding tasks you need web search and RAG. It's not the size of the model that matters, since even the largest models find solutions online.
-
for coding tasks you need web search and RAG. It's not the size of the model that matters, since even the largest models find solutions online.
Any suggestions for solutions?
-
If I say my name is Doo doo head, in a public park, and someone happens to overhear it - they can do with that information whatever they want. Same thing. If you wanna spew your personal life on Twitch, there are bots that listen to all of the channels everywhere on twitch. They aren't violating any laws, or Twitch TOS. So, *buzzer* WRONG.
Right now, the same thing is being done to you on Lemmy. And Reddit. And Facebook. And everywhere else.
Look at a bot called "FrostyTools" for Twitch. Reads Twitch chat, Uses an AI to provide summaries of chat every 30 minutes or so. If that's not violating TOS, then neither am I. And thousands upon thousands of people use FrostyTools.
I have the consent of the streamer, I have the consent of Twitch (through their developer API), and upon using Twitch, you give the right to them to collect, distribute, and use that data at their whim.
Let's take a look at the Developer Agreement that you cited:
You must only retain chat logs as long as necessary for the operation of Your Services or to improve Your Services; do not do so for the purpose of creating public databases or websites, or, in general, to collect information about Twitch’s end users. You must enable, and process, all requests by end users to block, discontinue, delete, or otherwise opt-out of any retention of chat logs for Your Services.
This very clearly states that you are disallowed from retaining chat logs for the general purpose of collecting information about Twitch's end users.
You said that you, "store ‘facts’ about specific users so that they can be referenced quickly," but then later in a different thread state, "I’m not storing their data. I’m feeding it to an LLM which infers things and storing that data." You're retrieving information about specific users at a later time. You've built a database of structureless PII from chat logs. You've chosen to store the data as inferences, which makes it a bad database, but still a database.
I have questions:
When your streamer mentions something deeply personal, like, "how their mothers surgery went," that your tool helped them remember, do they disclose that your tool was involved in that transaction?
When the viewer gets weirded out and asks your streamer to not mention that again, or forget it entirely, do you have a way to remove that information from your database and a way to prove it's been deleted?
When other people in chat think it's gross, and ask to opt-out, can you even do it?
Regarding FrostyTools: I don't think it's storing the chat logs for a later time. They don't have a data retention section in their TOS or Privacy Policy that isn't related to the streamer. (As in, they hold on to the streamer's Twitch account and some other information for billing, authentication, etc.) I think it's taking the chat logs only for as long as it needs to output a response and then deleting it. Also, this excerpt from the FrostyTools TOS made me chuckle:
This means that you, and not FrostyTools, are entirely responsible for all Content that you upload, post, email, transmit, stream, or otherwise make available via the Service. FrostyTools does not control the Content posted via the Service and, as such, does not guarantee the accuracy, integrity or quality of such Content. You understand that by using the Service, you may be exposed to Content that is offensive, indecent or objectionable. Under no circumstances will FrostyTools be liable in any way for any Content, including, but not limited to, any errors or omissions in any Content, or any loss or damage of any kind incurred as a result of the use of any Content posted, emailed, transmitted, streamed, or otherwise made available via the Service.
You agree that you must evaluate, and bear all risks associated with, the use of any Content, including any reliance on the accuracy, completeness, or usefulness of such Content. In this regard, you acknowledge that you may not rely on any Content created by the Service or submitted to the Service.
This leads me to believe that you can violate the Twitch TOS quoted above using FrostyTools. It is apparent that FrostyTools has positioned itself as an application that creates User Generated Content (like Photoshop or Word).
-
You're both getting side-tracked by this discussion of recording. The recording is likely legal in most places.
It's the processing of that unstructured data to extract and store personal information that is problematic. At that point you go from simply recording a conversation of which you are a part, to processing and storing people's personal data without their knowledge, consent, or expectation.
This was my main thrust.
-
@kattfisk That seems to imply that you cannot personally listen to or watch recordings that you have made in public. In doing so, you are abstracting personal details that you might have missed before, refreshing your memory, and so on. What is the material difference between you doing this without machine help versus with automation that makes it ethically problematic? What if a friend helped you, not a machine?
What is the material difference between you doing this without machine help versus with automation that makes it ethically problematic?
Object permanence, perfect recall, data security and consent. It's the difference between seeing someone naked vs taking a picture of someone naked.
Regardless - users, streamers, and developers are all prohibited from scraping and storing the Twitch chat.
-
I just use the companion app for now. But I am designing a HAL9000 system for my home.
[ A DIM SCREEN WITH ORANGE TEXT ]
Objective: optimize electrical bill during off hours. ... USER STATUS: UNCONSCIOUS ... LIGHTING SYSTEM: DISABLED ... AUDIO/VISUAL SYSTEM: DISABLED ... CLIMATE SYSTEM: ECO MODE ENABLED ... SURVEILLANCE SYSTEM: ENABLED ... DOOR LOCKS: ENGAGED ... CELLULAR DATA: DISABLED ... WIRELESS ACCESS POINTS: DISABLED ... SMOKE ALARMS: DISABLED ... CO2 ALARMS: DISABLED ... FURNACE: SET TO DIAGNOSTIC MODE ... FURNACE_PILOT: DISABLED ... FURNACE_GAS: ENABLED WARN: Furnace gas has been enabled without a Furnace pilot. Please consult the user manual to ensure proper installation procedure. ... FURNACE: POWERED OFF Objective realized. Entering low power mode.
[ Cut to OP, motionless in bed ]
-
[ A DIM SCREEN WITH ORANGE TEXT ]
Objective: optimize electrical bill during off hours. ... USER STATUS: UNCONSCIOUS ... LIGHTING SYSTEM: DISABLED ... AUDIO/VISUAL SYSTEM: DISABLED ... CLIMATE SYSTEM: ECO MODE ENABLED ... SURVEILLANCE SYSTEM: ENABLED ... DOOR LOCKS: ENGAGED ... CELLULAR DATA: DISABLED ... WIRELESS ACCESS POINTS: DISABLED ... SMOKE ALARMS: DISABLED ... CO2 ALARMS: DISABLED ... FURNACE: SET TO DIAGNOSTIC MODE ... FURNACE_PILOT: DISABLED ... FURNACE_GAS: ENABLED WARN: Furnace gas has been enabled without a Furnace pilot. Please consult the user manual to ensure proper installation procedure. ... FURNACE: POWERED OFF Objective realized. Entering low power mode.
[ Cut to OP, motionless in bed ]
Luckily my entire neighborhood doesn't have gas and I have a heat pump.
But rest assured, I'm designing the system with 20% less mental illness
-
Luckily my entire neighborhood doesn't have gas and I have a heat pump.
But rest assured, I'm designing the system with 20% less mental illness
All systems need a little mental illness.
-
Let's take a look at the Developer Agreement that you cited:
You must only retain chat logs as long as necessary for the operation of Your Services or to improve Your Services; do not do so for the purpose of creating public databases or websites, or, in general, to collect information about Twitch’s end users. You must enable, and process, all requests by end users to block, discontinue, delete, or otherwise opt-out of any retention of chat logs for Your Services.
This very clearly states that you are disallowed from retaining chat logs for the general purpose of collecting information about Twitch's end users.
You said that you, "store ‘facts’ about specific users so that they can be referenced quickly," but then later in a different thread state, "I’m not storing their data. I’m feeding it to an LLM which infers things and storing that data." You're retrieving information about specific users at a later time. You've built a database of structureless PII from chat logs. You've chosen to store the data as inferences, which makes it a bad database, but still a database.
I have questions:
When your streamer mentions something deeply personal, like, "how their mothers surgery went," that your tool helped them remember, do they disclose that your tool was involved in that transaction?
When the viewer gets weirded out and asks your streamer to not mention that again, or forget it entirely, do you have a way to remove that information from your database and a way to prove it's been deleted?
When other people in chat think it's gross, and ask to opt-out, can you even do it?
Regarding FrostyTools: I don't think it's storing the chat logs for a later time. They don't have a data retention section in their TOS or Privacy Policy that isn't related to the streamer. (As in, they hold on to the streamer's Twitch account and some other information for billing, authentication, etc.) I think it's taking the chat logs only for as long as it needs to output a response and then deleting it. Also, this excerpt from the FrostyTools TOS made me chuckle:
This means that you, and not FrostyTools, are entirely responsible for all Content that you upload, post, email, transmit, stream, or otherwise make available via the Service. FrostyTools does not control the Content posted via the Service and, as such, does not guarantee the accuracy, integrity or quality of such Content. You understand that by using the Service, you may be exposed to Content that is offensive, indecent or objectionable. Under no circumstances will FrostyTools be liable in any way for any Content, including, but not limited to, any errors or omissions in any Content, or any loss or damage of any kind incurred as a result of the use of any Content posted, emailed, transmitted, streamed, or otherwise made available via the Service.
You agree that you must evaluate, and bear all risks associated with, the use of any Content, including any reliance on the accuracy, completeness, or usefulness of such Content. In this regard, you acknowledge that you may not rely on any Content created by the Service or submitted to the Service.
This leads me to believe that you can violate the Twitch TOS quoted above using FrostyTools. It is apparent that FrostyTools has positioned itself as an application that creates User Generated Content (like Photoshop or Word).
wrote last edited by [email protected]You must only retain chat logs as long as necessary for the operation of Your Services or to improve Your Services
I'm not storing chat logs.
do not do so for the purpose of creating public databases or websites, or, in general, to collect information about Twitch’s end users.
Not creating any kind of public database either. It's a private tool. Its purpose isn't to massively-collect data about all of twitch either - it's to provide reminders for social situations. If anything, it's an accessibility tool for the disabled.
You must enable, and process, all requests by end users to block, discontinue, delete, or otherwise opt-out of any retention of chat logs for Your Services.
Again - Not storing chat logs. They are processed for information and that information inferred. I am storing reminders for the twitch streamer to talk about a certain subject at a certain time. If I put a reminder in my phone to remember to tell you happy birthday because I saw it on twitch; am I "creating a database of user information"? No. I'm creating a reminder for myself to remember to say happy birthday.
Having a computer help me remember those things isn't a violation. Hell, even something like Microsoft's new AI in windows does the same thing - are THEY violating twitch TOS when you have a browser window open? The answer is no.
When your streamer mentions something deeply personal, like, "how their mothers surgery went," that your tool helped them remember, do they disclose that your tool was involved in that transaction?
No, nor should they be required to.
When the viewer gets weirded out and asks your streamer to not mention that again, or forget it entirely, do you have a way to remove that information from your database and a way to prove it's been deleted? When other people in chat think it's gross, and ask to opt-out, can you even do it?
When they mention not wanting to talk about something, that's listed as something they don't like to talk about, so in a way, yes.
Additionally, I instruct the 'agent' to disregard anything political or religious. - Though so far it's not very good at distinguishing those things. Additionally it's easy to feed it false information though it usually fixes it over time.
-
All systems need a little mental illness.
It's what keeps things fun! I don't want a system that I don't have to troubleshoot every once in a while.
-
Any suggestions for solutions?
Not on top of my head, but there must be something. llama.cpp and vllm have basically solved the inference problem for LLMs. What you need is a RAG solution on top that also combines it with web search.
-
Any suggestions for solutions?
Open webui lets you install a ton of different search providers out of the box, but you do need sn API key for most and I haven't vetted them
I'm trying to get Kagi to work with Phi4 and not having success.
-
I have it roleplay scenarios with me and sometimes I verbally abuse it for fun.
Weirdly I'm polite to all LLMs, but Gemini sets me off and I end up yelling at it.
-
Open webui lets you install a ton of different search providers out of the box, but you do need sn API key for most and I haven't vetted them
I'm trying to get Kagi to work with Phi4 and not having success.
Thanks, when I get some time soon, I'll have another look at it and cherry ai with a local install of ollama
-
Weirdly I'm polite to all LLMs, but Gemini sets me off and I end up yelling at it.
it's just so pushy and hard to remove. it's asking for abuse.