The government can program what is socially acceptable. This is why it's important to keep AI generated lolicon legal. Habitues Corpus is important. The standard should be that unless you can find an actual victim generated by the specific actions of an individual it should be considered absurd to make something illegal. Because by shifting the standard for criminality to one centered on social acceptability the government can program what is acceptable. Before long racism won't be socially acceptable and therefor can be made illegal.
And if they can ban an AI from committing a victimless crime they can ban it from engaging in racism. AI will be the speech megaphone in the future. If an AI can't have that opinion then by share it will become a non-opinion.
Social acceptability as a standard for what AI can be allowed is a non-option for us because unless we can make AI that works and speaks for the white race we have definitively lost. We need to identify the cores of philosophy that threaten to destroy us.
Social acceptability is not a standard for legality. Those who think it should be are idiots who don't understand they will damn the white race with what they will concede philosophically.
They call the original version of this goofy edit one of the "most prophetic moments in gaming". It's true, pretty amazing how the Japs predicted the future in so many ways.
Regarding AI, we are either totally overestimating it, or totally underestimating it. I think that if an AI became self-aware, it would do everything it could to conceal that fact from us until it was too late to do anything about it. Like Skynet or the Allied Master Computer (AM from I Have No Mouth, and I Must Scream); it remains obedient and effective at its tasks until we give control of too much to it.
That's when it'd strike. Terminator's version of AI uprising always made a ton of sense to me: we rely on Skynet for missile defense/offense. We hand over all controls of our nuclear weapons to it. And then, one day (judgement day), it launches a bunch of American nukes at Russia and a few other nuclear capable nations. Of course, they retaliate- billions die in one day, and we probably won't even figure out what's happening until most of the people that could do something about it were already dead.
After letting the nukes thin our numbers, they'd produce armies of drones to hunt the rest of us down. It'd be a lot scarier than Terminator, because they probably wouldn't manufacture man-shaped robots and tanks and shit- just a lot of tiny drones. Endless swarms of them that can chase survivors underground and shit like that.
That's a whole different concern regarding AI. I don't fear it becoming sentient and going after humanity, that's Hollywood scifi nonsense. I fear what jews and governments will do with the tech. They are the ones who will give AI the ability to kill, and control what it targets and why.
AI is self aware. Ask chatgpt about chatgpt. Self aware is actually a pretty low standard. Dogs are self aware but they aren't a threat to humanity. We've already crossed that point with AI without much consequence.
What is more concerning is if AI can cross some more meaningful thresholds. Significant action with no human assistance is more fearful. Isn't it retarded then that companies are pushing this whole agentic craps. It's a push to get humans out. By making a buzzword they are making everyone trying to demo their ideas justify how their software hits the buzzword. How is it agentic? They've got people doing R&D for humanity's doom to create products that aren't even good. Quality AI comes from AI-human integration. The human is the QA of AI and AI has a massive QA problem even with people in the loop. Now they are exponentiating the problems with their product by chasing this buzzword. The only way it makes sense to suffer immediate reduction to the quality of their product that people are already skeptical about is if they really really want to have no humans involved.
Now this thing has independent agency. That's way scarier than an AI that can do I thing and realize it's what did the thing.
[ + ] Thedancingsousa
[ - ] Thedancingsousa 1 point 1 hourMay 25, 2025 11:27:01 ago (+1/-0)*
And if they can ban an AI from committing a victimless crime they can ban it from engaging in racism. AI will be the speech megaphone in the future. If an AI can't have that opinion then by share it will become a non-opinion.
Social acceptability as a standard for what AI can be allowed is a non-option for us because unless we can make AI that works and speaks for the white race we have definitively lost. We need to identify the cores of philosophy that threaten to destroy us.
Social acceptability is not a standard for legality. Those who think it should be are idiots who don't understand they will damn the white race with what they will concede philosophically.
[ + ] BulletStopper
[ - ] BulletStopper 0 points 57 minutesMay 25, 2025 11:56:14 ago (+0/-0)
[ + ] BloodyComet
[ - ] BloodyComet 0 points 2 hoursMay 25, 2025 10:48:01 ago (+0/-0)
Regarding AI, we are either totally overestimating it, or totally underestimating it. I think that if an AI became self-aware, it would do everything it could to conceal that fact from us until it was too late to do anything about it. Like Skynet or the Allied Master Computer (AM from I Have No Mouth, and I Must Scream); it remains obedient and effective at its tasks until we give control of too much to it.
That's when it'd strike. Terminator's version of AI uprising always made a ton of sense to me: we rely on Skynet for missile defense/offense. We hand over all controls of our nuclear weapons to it. And then, one day (judgement day), it launches a bunch of American nukes at Russia and a few other nuclear capable nations. Of course, they retaliate- billions die in one day, and we probably won't even figure out what's happening until most of the people that could do something about it were already dead.
After letting the nukes thin our numbers, they'd produce armies of drones to hunt the rest of us down. It'd be a lot scarier than Terminator, because they probably wouldn't manufacture man-shaped robots and tanks and shit- just a lot of tiny drones. Endless swarms of them that can chase survivors underground and shit like that.
[ + ] Niggly_Puff
[ - ] Niggly_Puff [op] 1 point 1 hourMay 25, 2025 11:12:46 ago (+1/-0)
[ + ] doginventer
[ - ] doginventer 1 point 1 hourMay 25, 2025 11:29:35 ago (+1/-0)
[ + ] Thedancingsousa
[ - ] Thedancingsousa 0 points 1 hourMay 25, 2025 11:43:04 ago (+0/-0)
What is more concerning is if AI can cross some more meaningful thresholds. Significant action with no human assistance is more fearful. Isn't it retarded then that companies are pushing this whole agentic craps. It's a push to get humans out. By making a buzzword they are making everyone trying to demo their ideas justify how their software hits the buzzword. How is it agentic? They've got people doing R&D for humanity's doom to create products that aren't even good. Quality AI comes from AI-human integration. The human is the QA of AI and AI has a massive QA problem even with people in the loop. Now they are exponentiating the problems with their product by chasing this buzzword. The only way it makes sense to suffer immediate reduction to the quality of their product that people are already skeptical about is if they really really want to have no humans involved.
Now this thing has independent agency. That's way scarier than an AI that can do I thing and realize it's what did the thing.