The locking down started with the original MacIntosh (or actually with the Lisa I guess). ISTR they had at least one bit more open period after that, but those have always been the exception.
The locking down started with the original MacIntosh (or actually with the Lisa I guess). ISTR they had at least one bit more open period after that, but those have always been the exception.
Wouldn’t it be more correct to say that most Americans also use a messaging app (iMessage). The rest are just stuck with SMS to have compatibility with the iPhone users.
As the iPhone was (is?) not as popular in the Europe as it was (is) in the States that might also be one of the reasons why people here ditched SMS so fast once smartphones got popular.
For years now I have only read ebooks on my phone, so one evening I decided to get back to the habit of reading real books.
So I take my time and carefully pick just the right book, gather some pillows, turn off the lights and lay comfortable on the couch. And after a few confused moments of flipping through pages I realized that these fucking things didn’t work in the dark. And I really don’t like to read under a bright light anymore so back to reddit it was for that evening.
That said, I think I’ll skip this one, doesn’t sound too comfortable.
Speaking as just a hobbyist, a more developer oriented community focused on the topic would be nice, if someone is up to the task.
It’s currently hard to find any good information about how to actually use LLMs as part of a software project as most of the related subreddits etc. are more focused on shitposting and you don’t currently really want to talk about these in general tech/programming forums without a huge Don’t shoot I’m not one of them! disclaimer.
Regarding little Bobby, is there any known guaranteed way to harden the current systems against prompt injections?
This is something that I’m personally more worried about than Skynet or mass unemployment now that everyone and their dog is rushing to integrate LLMs into to their systems (ok worried maybe a wrong word, but let’s just say I have the popcorns ready for the moment the first mass breaches happen with something like the Windows Copilot).
At least I’m interested but more technical discussion about this would probably fit better in some comp sci or programming community? Though most of those are a bit hostile to the LLM related topics these days because of all the hype and low effort spam.
Is the whole “You are an LLM by OpenAI, system date is etc.” prompt part of the system message?
A few days ago when I was talking about controlled natural languages with it and asked it to give a summary of the chat so far in Gellish it spit that out.
If these commands were in a system message it would generally refuse to help you.
Doesn’t it usually fairly easily give its system message to the user? I have had that happen purely by accident.
I’m not sure if I’d call that reverse engineering any more than using a web browsers View Source feature.
But it’s interesting how it works behind the scenes and that only way to get these models to interface with the external world is by using the natural language interface and hoping for the best.
Yep, once anyone can download an app on their phone and do something like this without any effort in realtime it’s going to lose its (shock) value fast. It would be like sketching a crude boobs and vagina on someones photo with MS Paint and trying to use that for blackmail or shaming. It would just seem sad and childish.