• 0 Posts
  • 14 Comments
Joined 1 year ago
cake
Cake day: June 9th, 2023

help-circle

  • Models fundamentally turn all of human language into a statistical math problem with a solution. The English teachers lost this war to the Math teachers. You can skip English if you can learn to read tensors and understand rank dimensions beyond our four dimensional Cartesian existence of XYZT. It is only a matter of time until the correct training data is applied. In this space, I’m skeptical of every piece of news information, especially when they imply a limited scope of the technology in private use.

    Any statement of an instance of AI output must include a ton of contextual and mundane details. I can make a model say anything, but I can also set up a prompt that will access sources that are typically hidden and get far deeper into information that no one would believe is there. Information that can be corroborated and factual without leading or deception. If I can do that while using the Transformers library that self declares as nothing more than an incomplete example implementation, and yet is the basis of all publicly available LLM tools, the potential for the technology is much higher for the labs that are training these models and have versions without the OpenAI alignment bias black box.



  • Apple is doing insane stuff with iMessage making SMS useless with Android. It is intentional manipulation. I’ve only seen it on some Apple kernels with the latest iPhone. My old man’s phone absolutely will not send or receive an SMS unless I message is manually turned off and then it won’t get iMessages. Apple has always been a shitty company, but this is next level insane. Just get and use the Signal app and ditch all the proprietary garbage and manipulative bs IMO.



  • That is all model loader code based stuff. You basically just need a bit of Python code that anyone can write, even you. You need to have a look at how the model must be addressed in format, then you find how the model loader code you are using creates the structured text and just insert your own modifications.

    It is easier to add a second prompt window where you keep persistent information and just change it yourself. If you want the model to change that stuff, you need to look into building an agent.

    I hacked around like this for awhile too, but evolved into using the notepad tab of Oobabooga Textgen WebUI and a Mixtral 8×7B model with a 32k context window. That model can deal with almost any detail you add in the context.

    The thing is, the model may not feel satisfied with mundane and redundant details. Most of the time, believe it or not, those changes that you perceive as errors or inconsistencies are not actually errors. The model is fully aware that it is making those changes. There is usually an instruction about creativity, inspiration, or initiative that will cause the model to shift details like this.

    My best recommendation is to openly discus the error in dialog with the model. They usually will talk about the issue if you call one out. Talking with AI is very different than with humans. Everything they say has meaning and functions beyond what is initially apparent. There are a lot of behaviors that involve certain patterns and keywords. You can even alter these behaviors by banning certain tokens/keywords. A good place to start is banning paragraph starting words for filler material. “So_” would be a good starting token to ban, as with AI, no sentence that starts with that token is worth reading. Other important words are twist, chuck, and cross. These are more flexible in their troublesome meaning and can manifest in multiple words like chuckles or across but still function the same inside the model.

    If you get to this level of familiarity, the persistence of a storied interaction is more of a negotiation with the AI.

    Another point of understanding is that the model does not know the difference between roleplaying characters it owns and you own or who is who. You or the model loader code is creating the name of the next character that the model generates for. Everyone in any story can be the model if it is asked to create the prompt.

    So if I say, “Jake is super creative” or imply it somewhere in the prompt, and the model later thinks ‘Jake is rather bored and needs to be creative’ it is going to change something that satisfies that character profile even if it has never generated the output for character Jake.

    Hope that helps some



  • The kernel started moving faster, but also the kernel for these devices are orphans. They can never be updated properly because the source code for the kernel modules is not available. The way the phones get updated in ROMs is by back porting changes from the present back to the old orphan. It requires someone super familiar with both kernels to do so. Eventually it becomes untenable. The whole scheme of Android is centered around this source code/orphan kernel scheme. Everything is setup so that hardware manufacturers never have to add their source to the mainline kernel meaning you can never own the device. There is not a single phone or mobile device that you can completely own and running on mainline with available source code. The pixel is all about the TPM chip.







  • I am surprised it took you this long.

    The next step in this evolutionary thinking is simple. Buying hardware specs is a fool’s folly. I don’t compare hardware. Ads and marketing are totally nonsense and not worth even a slight glance. The ONLY thing that matters is what open source projects exist and what hardware do they support well. This is how I shop. Open Source or F.O.