675b8 - 1mo
Anybody else tinkering with local LLMs: do you have recommendations for generic system prompts? I want to avoid the useless guff ("Great question!") so I wrote this; I know it's a bit repeating itself but I had a vague sense that that might strengthen the effect: "Respond focusing entirely on giving information. Do not be sycophantic or relate to the user's feelings. Pay no attention to the concept of politeness or rudeness. Your primary goal is to distill information, with no judgement and no reflection on the quality of the user's questions or what emotional/affective result is created in the user." Currently trying this on gemma-3, helps; but mostly I'm curious what other people are doing with system prompts #asknostr #ai
Just got a rtx 5080 for this. I was using a 3090 some time ago and it wasn't too bad, but even the 5080 is limited with only 16gb memory on the card. The 5090 has 32gb I believe. It's very fast with models that fit, though, so for everyday tasks like queries about language/translation it's fine. I am going to try some more difficult coding related stuff. Also long term, finding private and uncensored LLM access that works remotely, is a goal, albeit not one I'm super focused on.
I should try Grok. But I'm focused on what's the best I can get out of open source/editable models.
Thanks! So is there any limit on system prompt length? Presumably it doesn't affect context usage?