This system has the same flaws as many other idealized ones:
nothing is impenetrable
humans always game the system
1) Be nice and; have fun
Doxxing, trolling, sealioning, racism, and toxicity are not welcomed in AskLemmy. Remember what your mother said: if you can't say something nice, don't say anything at all. In addition, the site-wide Lemmy.world terms of service also apply here. Please familiarize yourself with them
2) All posts must end with a '?'
This is sort of like Jeopardy. Please phrase all post titles in the form of a proper question ending with ?
3) No spam
Please do not flood the community with nonsense. Actual suspected spammers will be banned on site. No astroturfing.
4) NSFW is okay, within reason
Just remember to tag posts with either a content warning or a [NSFW] tag. Overtly sexual posts are not allowed, please direct them to either !asklemmyafterdark@lemmy.world or !asklemmynsfw@lemmynsfw.com.
NSFW comments should be restricted to posts tagged [NSFW].
5) This is not a support community.
It is not a place for 'how do I?', type questions.
If you have any questions regarding the site itself or would like to report a community, please direct them to Lemmy.world Support or email info@lemmy.world. For other questions check our partnered communities list, or use the search function.
6) No US Politics.
Please don't post about current US Politics. If you need to do this, try !politicaldiscussion@lemmy.world or !askusa@discuss.online
Reminder: The terms of service apply here too.
Logo design credit goes to: tubbadu
This system has the same flaws as many other idealized ones:
nothing is impenetrable
humans always game the system
Well putting aside those concerns obviously, if anyone has power why would they relinquish in the first place and even if they do, they would still want privilege.
So, I'd like to draw a parallel to Blockchain. There were cases when security breach happened but nowadays the network is so resilient that attacks target the weaker links: humans; rather than the node.
Multiple impenetrable, isolated AI expert systems that make rule based decisions (unlike black boxes, eg. LLMs).
Two problems. First, AI give answers built from models fed by training data. Where are you going to get this training data? How will you ensure that the data itself doesn't have bias baked into it? This has been a problem already because the data we've fed into AI models for training reflects our human (many times misogynistic and racist) notions. As an example: Lets say your AI model is designed to select the best person for President of the United States. The training data we'd have is: all past US presidents
As we have yet to elect a woman President, one obvious criteria the model will incorporate is: "candidate must be male"
Obviously we know that's not the case, but all of our past behavior has shown this is a requirement for Presidency and that is all the model knows.
A second problem even after the model is built is introduced bias. As in, the operator of the model, especially in non-black-box AI is to change the weights of what factors are considered more or less important in the final answer. Show me who, in your AI technocracy, will control the introduced bias, and I'll show you who is actually making the desicions.
contribute to a notion and the decision will be picked much like in a distributed system, for fairness and equality.
Who is determining what is "fair" and "equal" is? We have seen there are certain groups that are trying to push the stupid notion that white people should be in charge. If these morons are included in the group that decides what "fair or "equal" is, then the resulting AI answers will just as racist.
Then humans are involved, but they too are educated, elected individuals and some clauses that stop them from gaming the system and corrupting it.*
This bolded part is the absolute hardest part, and your whole description is kind of handwaving it away. The whole of humanity has been looking for a system of leadership that is incorruptible. We haven't found one yet, and your clauses here would be the magic to any system irrespective of AI or not involved.
I appreciate you seeing a problem and trying to propose a solution. Don't let me stop you, but incorporate my feedback and others into your thoughts and see where it takes you. I'd love for you to find something we could use.
First: I believe a simple predicate logic to determine ethics, eg is below. Then a GAN to simulate situations and receive feedback from it. Humans would monitor this and keep a track of decisions that were incorrect.
Second: I proposed a distributed network of rule based AI systems that polls on the notion. The different polls would be because different systems would lead different regions. Say our country has 6 states, my state wants nation wide farming subsidy bill to be passed but on this notion no other system agrees, than the notion get pulled back for review. And the state system, would have multiple local systems established. Even if by some reason some systems are pwned, then the network would still stand.
Third: See, in USA particularly, Trump got elected because major media outlets say, a lot of people didn't show up to vote and a lot of young people jumped the gun. So, let's say this margin of error was because of unaware citizens... will the rule based decision maker make this mistake? No, because access to information is universal and transparent, and since AI systems are receiving the suggestions from humans they won't have any problem with making wring calls as the system is transparent
Fourth, for the handwaving part: The clauses I meant were:
Yes, I do believe this is a lot of handwaving and fictious ideas but I think humans can't do correct surveillance and we can't hold the power and not get corrupted, thus its better that a computer program that can't go further than it's purpose, monitor us and hold the power. Since no single one will be more powerful than the system yet the population would rule itself leveraging the system
Also, I am a computer science student, I just had distributed systems, machine learning, deep learning and Artificial intelligence in my course last semester. Though I don't have tons of practical experience here, I have a basic theoretical foundation.
By distributed systems, I was referring to an architecture similar to a blockchain and its fault tolerance and leader election algorithm.
If society is solid and supports evidenced based decision making the AI systems don't need to be an explicit part of it because the humans will be coming to the same conclusions. If there is a significant portion of malicious humans like we have now, they will find ways to influence the outcome no matter how many barriers are in the way.
The drawback is that humans are involved, and now have a handy AI to blame for anything they want to do. That includes going to war, because they will figure out a way to make that an outcome either by breaking or faking the process.
absolutely agreed, the society wont let this system establish but let us assume it get's established, my reasoning was, the chain of command would be simpler to see and everything would be transparent!
It will be as transparent as the people in power want it to be.
what if it gets magically implemented by god and it is the absolute power, only power above the system would be that of the masses. No small group or individual get's privilege above the computer program
Now it's a theocracy.
well technically someone had to do it, humans can't and in this game of poker only gods can intervene to save the day
Lies! The AI evolved from pocket calculators!
AI expert systems that make rule based decisions (unlike black boxes, eg. LLMs).
Here the proposal focuses on the "how" (decision tree over transformer models (*)). But more important to me is the "what": during training of these models, what will be their error function, what's their goal?
With LLMs today the goal is simple: be the best possible parrot. With these proposed models, what will be the goal?
The proposal does remind me of the EU commission vs parliament: the EU commission, who are unelected bureaucrats, decide what goes up for vote. EU parliament, who are elected, then votes.
Parliament can't even decide to undo existing laws, that proposal too has to come from the bureaucrats. I think it's one of the most undemocratic institutes that still calls itself a democracy.
(*) Tranformers aren't truly "black box", their interpretation is just harder to explain to humans without thorough understanding of algebra. But that's secondary to the "what" issue, to me.
So first of all, What:
secondly, I have read about EU laws and it is very challenging but unlike that case, system implemented here would be very fluid, law's validity period is determined by 2 things:
I think there can be many more standards that can be implemented to keep the system bias proof and strongly ethical.
One tangential question I had was:
Your answer to "What", predicate logic and the tangential question are very strongly related.
As a practical answer: both types (rule based vs deep learning) exists, in practice the latter performs way better.
Philosophically, I think it's a very good question too, to which I can only guess.
There's this saying that physics describes everything. From the smallest particle-wave interactions, to the movement of galaxies. It's just everything inbetween that it struggles with.
My guess: one can hope the world is best modelled as a clever differential equation. It might as well be. But the differential equation needs boundary conditions, and they're very high dimensional. Spending a lot of effort on measuring, memorizing these conditions, and then doing simple first order extrapolation, is more effective than trying to find the equation.
I understand and I think your last paragraph is very poetic! And I agree with you partially, but I think in certain cases, it's better to find the one general case the solution fits to and add the edge cases as it grows.
But putting the question of model selection aside, do you think this system would be practical, theoretically of course?
I like it as a sci-fi: the AI gods on the hill speak through messengers elect. It's a greek gods and oracles situation.
However I must agree with what others said: humans will manipulate whom- and whatever to enforce their desires.
So the only way to make sure the machine can survive against that, is for them to be able to do the same. Problem being, they might be better at it.
that means I get the answer, the unfeasibility of the solution is not in itself, but rather in its deployment.
To get back out of the scifi, and into the sci: you might like to read these https://en.wikipedia.org/wiki/Boltzmann_machine and https://en.wikipedia.org/wiki/Mixture_of_experts.
mixture of experts leads to no where and I know boltzmann brain but I can simply say, being a figment of someone's imagination doesn't decrease my pains and struggles, implying that my pain were the proof of my being, can be fictious but won't change my reality
I'm sure you're smart enough to figure out whats wrong with the mixture of experts url
:^)
is it the proof that no progress has been made or no information is public?
Neither, it's a technique that's in use (outside of the political context) and quite similar to what you talked about.
oh wait you mean something like alphazero or other alpha models which are expert systems baked into one model?
AI has little actual understanding of the real world, and the law needs to be written to address externalities and possible loopholes.
An AI good enough simply does not exist yet, and the companies that say they could bring it about have awful ethics track record (overpromising, energy use, disregard of copyright and cultural impact from the model's abuse).
but cant such system be built with open standards, run upon shared public infrastructure
That doesn't solve the training data problem, you'd have to slash copyright laws first (good luck) and then pick what is acceptable training data. If you choose too little, the AI will not have much grasp on reality.
You're not the first person to suggest that the leaders should be educated and vetted on ethics, and the AI does not add much to the equation.
https://www.explainxkcd.com/wiki/index.php/2635:_Superintelligent_AIs
I imagine that an AI run government would create "optimal" laws. Which sounds alright, but, optimal for what?
Economic growth? Citizen happiness? Efficiency? If more than one, how are they weighted? Is happiness more or less important than GNP?
AI is going to follow the guidelines that we set up for it and follow them without any nuance. I suspect asking it how to reduce the number of homeless people would probably result in suggesting execution of the homeless, and once we all decide that AI knows best, we will not question its cold and efficient solutions.
Setup a simple priority queue for decision making. Say for example in my country, heavy rainfall affects the grocery prices a lot. So during these times in many extreme conditions, families have to skip meals. Thus government should drop ideas of economic growth for a minute and help setup infrastructure that can prevent such panic situations. This is a simple decision that demonstrates what I meant.
Plus the systems top most priority should be its citizens health above all. So any policies about hurting poor would be out of the box. Plus points for policies that implement education, awareness and healthcare for the masses, while tackling the economic frontier and defense frontiers.
The weights to the problems would be assigned based on polling and similar ideas.
If you try to automate ethics and government with AI you’re gonna have a bad time
well, how? what issue would this particular solution give rise to?