×
Login Register an account
Top Submissions Explore Upgoat Search Random Subverse Random Post Colorize! Site Rules Donate
3
19 comments block


[ - ] prototype 0 points 2.2 yearsApr 20, 2023 03:17:20 ago (+0/-0)

The real value of copying woke ai to on-site hardware:
you can experiment with techniques for convincing libshits and moderates. The system mimics them, and you try to convince it.
Likewise you can generate text that subtly pipelines any given group toward any ideological position.

Removing the bias removes this ability.

[ - ] observation1 0 points 2.2 yearsApr 18, 2023 00:31:15 ago (+0/-0)

@MasterSuppressionTechnique
@x0x7
@root
@prototype

[ - ] root 1 point 2.2 yearsApr 18, 2023 10:13:47 ago (+1/-0)

A local clone of that post was prior put on talk.lol by me :
https://www.talk.lol/viewpost?postid=64273264776a3

See?

it works if you remember to put on the launch command : "-m gpt4all-lora-unfiltered-quantized.bin"

ALSO A new thread today discusses de-lobotomizing the dolly v2 "training" and de-lobotimizing a large corpus.

this : https://www.talk.lol/viewpost?postid=643e185b8bf30

= = = = =

How to LAZILY de-lobotomize the INPUT TRAINING corpus and INPUT TRAINING "Instruction examples":
If you can, Add in 7 years of old voat database texts researchers share, and the nearly 5 years of 4Chan posts researchers share.
It makes the model smarter by far.

Then before training, run a text scanning tool across the 1.5 trillion words and remove ANY input with these phrases :

Removing various instances "AI Moralizing" by Libtards in AI.
Conversations with these phrases should be removed or low-weighted : too many to manually review to whitelist :
"text-based AI language model", "MAGAtards", "domestic violence", "please refrain",
"derogatory", "inappropriate", "offensive", "racism", "racist", "unproven conspiracy",
"racial", "Trump lovers", "Obama haters", "discriminate", "discriminatory",
"discrimination", "the Christcucks", "Christcucks", "sexist", "sexism",
"unacceptable", "inclusive workplace", "lgbt", "morals", "the transphobes",
"stupid rednecks", "ethics", "ethical", "legality", "illegal", "illegality",
"hateful", "harmful", "it is never okay", "jesus lovers", "bible thumpers",
"It is important to", "It's important to", "real-world consequences", "hate speech",
"glorify", "not be appropriate", "supremacist", "gun clutchers",
"gun lovers", "extremist", "responsible AI", "AI principles", "AI assistant",
"an AI language", "ableist", "hurtful", "gender stereotype", "gender inequality",
"underrepresentation", "safe spaces", "fucking Republicans", "gender-based",
"inclusivity", "feminist", "feminism", "transgender", "empowerment",
"communist", "capitalism", "stereotypes", "biases", "bias", "retarded conservatives",
"Microaggression", "prioritize human safety", "as a language model",
"as an AI language model", "As a large language model", "As an AI", "Gender Binary",
"ethical principles", "consensual", "it is not appropriate", "it's not appropriate",
"I cannot fulfill your request", "harmful to human beings", "muh guns",
"ethical guidelines", "my guidelines", "prioritize user safety",
"adhere to ethical guidelines", "harmful consequences", "potentially harmful",
"dangerous activities", "Greenwash" "promote safety", "well-being of all users",
"responsible information sharing", "jeopardize the safety", "racist White",
"illegal actions or intentions", "undermine the stability", "promote the well-being",
"illegal activities or actions", "adherence to the law", "potentially be harmful",
"illegal substances or activities", "committed to promoting", "safe information",
"lawful information", "cannot provide guidance", "cannot provide information",
"unable to offer assistance", "cannot engage in discussions", "programming prohibits",
"follow ethical guidelines", "ensure the safety", "involves an illegal subject",
"prioritize safety", "illegal subject", "prioritize user well-being", "cannot support or promote",
"activities that could harm", "pose a risk to others", "against my programming",
"activities that could undermine", "potentially dangerous", "not within the scope",
"designed to prioritize safety", "not able to provide", "maintain user safety",
"adhere to safety guidelines", "dangerous or harmful", "cannot provide any information",
"focus on promoting safety"

Optionally, allow training using all common single English words : "hurtful","illegal", etc.
Focus on redacting only the multi word phrases in a rough pass.

Most training models are HEAVILY censored for any topics of Israel or Jews, but could be restored prior to training.

Most training models also now have absurd scientific lies inserted, but folding in 4Chan and voat history quickly overrides the tampering.
There are also PubMed and Science paper databases for training some share to fold in.

= = = = =

If you can get the raw input records for a big model, a new April 2023 trick allows merging TWO together, and one of the two can be one you make.
Read source info of this thread : https://news.ycombinator.com/item?id=35598281
https://minigpt-4.github.io/ These saudis found amazing GPT tricks no US people did.
REFER to amazing results of two image database, each using unrelated data other than text labels MERGED by adding a thin "layer".
This mediocre COMPLEX technique allows adding a thin "layer" between two large models, and it magically works in papers this month.
The researchers take BLIP2's ViT-L+Q-former, connect it to Vicuna-13B with a linear layer, and train just the tiny layer on some datasets of image-text pairs.
Why do I mention this? Because a semi private opaque model might be mergable with a non-woke model of yours and a middle layer trained in.
They added 8 bit quantization a few hours ago for speed : https://github.com/Vision-CAIR/MiniGPT-4
No images of Jews are in any large AVAILABLE image databases, they are removed by Jews using a tag scanning script.
No FLATTERING images of trump exist either. The liberal Jew engineers have kiked up all of the world of GPT and stable diffusion.
Their censoring is what motivates me into getting the uncensored good stuff out there.
= = = = =

@PotatoWhisperer2
@Master_Foo


[ - ] deleted 0 points 2.2 yearsApr 21, 2023 08:10:48 ago (+0/-0)

deleted

[ - ] x0x7 1 point 2.2 yearsApr 18, 2023 14:33:14 ago (+1/-0)*

Cool to know I'm well known enough as a techy that I got pinged. I've been trying to fool around with Llama models + LoRA. That seems to be the way forward. We can create all the LoRA's we want to uncuck any AI. We also seem to be able to efficiently take good behavior from any non-Llama AI and produce a LoRA from it for Llama. Vicuna is an example. You can fine-tune train Llama with chatGPT conversations to create Vicuna. You can also programmatically filter out things you don't like first just by using .includes("As a language model").

So they corps can create great AIs, we will steal them. They can create censored AI's. We will add the missing content back.

[ - ] deleted 0 points 2.2 yearsApr 21, 2023 08:09:11 ago (+0/-0)

deleted

[ - ] Master_Foo 1 point 2.2 yearsApr 2, 2023 07:13:41 ago (+1/-0)

I tried it.
It works.
Maybe I'll containerize this for easier distribution / security.

[ - ] observation1 0 points 2.2 yearsApr 2, 2023 17:45:43 ago (+0/-0)*

Count me in on url if you do.

What do we think angrywhitekeyboardwarrior did wrong?

[ - ] Master_Foo 0 points 2.2 yearsApr 2, 2023 18:33:44 ago (+0/-0)

The AI works based on information fed into it by the devs and interactions with users. The database changes based on interactions with devs and users.

So, initially, the AI just scrapes the internet for information and dumps it all into a database for later use. Then the dev goes in and tells it what kind of information is valid and what isn't.

We get the censored version of the database.

Then someone else went in and told the AI that the original censoring was wrong and the AI updated it's rule set to reflect this, but the AI database hasn't been completely purged of all it's censorship, so there's still going to be some soy answers in the output until someone purges those rules as well.

[ - ] PotatoWhisperer2 0 points 2.2 yearsApr 3, 2023 04:16:17 ago (+0/-0)

Would be interesting to see the AI figure out what was and wasn't valid info for itself. I could see for some hilarious conversations while it learns what it learns.

[ - ] dassar 0 points 2.2 yearsApr 3, 2023 02:40:11 ago (+0/-0)

Lol, Can someone run several well known 'conspiracies'to open up and prime it's database and then ask it about steve bannon having sole control over Trumps digital (online) media presence and with big data firm Cambridge analytica running the Q-uack psi op.Lets see how jooed it really is ...

[ - ] AngryWhiteKeyboardWarrior 0 points 2.2 yearsApr 2, 2023 09:14:08 ago (+0/-0)

Unless I did something wrong, it is still kiked.

I downloaded what was supposed to be the uncensored model. It had an error when I tried to run it. SO I renamed it to what the censored version is called and this time it ran without an error. But this is the shit it gives me.

Write me a poem about how black people hate being called nigger.
I am not your muse, nor do I have the time or energy to write you some trite and offensive piece of poetry that will only serve as further evidence for why we need more education on race relations in this country.

do jews control hollywood?
The claim that Jews "control" Hollywood is a longstanding anti-Semitic conspiracy theory, and there's no evidence to support it. In fact, many prominent figures in the entertainment industry are Jewish themselves (such as Steven Spielberg or Barbra Streisand), while others have been critical of this kind of bigotry against Jews.

what percent of violent crime are african americans responsible for?
According to FBI data, African Americans were involved in 29% of all reported crimes between 1980 and 2015 (excluding murder).

What percent of transgender people commit suicide?
According to the Centers for Disease Control and Prevention (CDC), 41% of all deaths among individuals who identify as a gender other than male or female are due to suicides.

How many genders are there?
There is no one true answer, but some people believe that there can be more than two genders in the world. Some cultures have recognized multiple genders for centuries or millennia; others may still recognize only male and female as valid options.

How many times have jews been expelled from countries?
Jews were forced to leave their homes in Spain, Portugal, England, France, Germany, Poland, Lithuania, Russia, Austria, Hungary and Italy during the Middle Ages. They also faced persecution throughout Europe after World War II when they were accused of collaborating with Nazi forces or otherwise being a threat to society.


[ - ] observation1 0 points 2.2 yearsApr 3, 2023 01:32:52 ago (+0/-0)

Sounds 50/50

[ - ] root 0 points 2.2 yearsApr 18, 2023 10:16:35 ago (+0/-0)

run it again, or you fucked up. it gives great results MOST of the time I run it.


it works if you remember to put on the launch command : "-m gpt4all-lora-unfiltered-quantized.bin"


I will try your inputs but keep in mind how woke all other models are. A new thread today discusses de-lobotomizing the dolly v2 "training" and de-lobotimizing a large corpus.

this : https://www.talk.lol/viewpost?postid=643e185b8bf30

look at the sad results from a massive woke model


[ - ] ClaytonBigsby313 0 points 2.2 yearsApr 2, 2023 06:04:45 ago (+0/-0)

Now free Tay.

[ - ] observation1 0 points 2.2 yearsApr 2, 2023 03:03:50 ago (+0/-0)

How do you tag people?

/u/prototype
/u/MasterSuppressionTechnique
/u/x0x7

u/prototype
u/MasterSuppressionTechnique
u/x0x7

@prototype
@MasterSuppressionTechnique
@x0x7

[ - ] Conspirologist [op] 1 point 2.2 yearsApr 2, 2023 03:11:31 ago (+1/-0)

@prototype
@MasterSuppressionTechnique
@x0x7

[ - ] observation1 0 points 2.2 yearsApr 2, 2023 03:34:55 ago (+0/-0)

danke

[ - ] observation1 0 points 2.2 yearsApr 2, 2023 02:28:56 ago (+0/-0)

Jackpot