this post was submitted on 04 May 2025
92 points (79.5% liked)
Technology
69772 readers
4546 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Humans aren't innately good at math. I wouldn't have been able to prove the statement without looking things up. I certainly would not be able to come up with the Peano Axioms, or anything comparable, on my own. Most people, even educated people, probably wouldn't understand what there is to prove. Actually, I'm not sure if I do.
It's not clear why such deficiencies among humans do not argue against human consciousness.
That's dubious. LLMs are trained on more text than a human ever sees, but humans are trained on data from several senses. I guess it's not entirely clear how much data that is, but it's a lot and very high quality. Humans are trained on that sense data and not on text. Humans read text and may learn from it.
What might an operational definition look like?
Just because you can't make a mathematical proof doesn't mean you don't understand the very simple truth of the statement.
I think if I could describe that, I might actually have solved the problem of strong AI.
You are asking unreasonable questions.
If I can't prove it, I don't know how I can claim to understand it.
It's axiomatic that equality is symmetric. It's also axiomatic that 1+1=2. There is not a whole lot to understand. I have memorized that. Actually, having now thought about this for a bit, I think I can prove it.
What makes the difference between a human learning these things and an AI being trained for them?
Then how will you know the difference between strong AI and not-strong AI?
I've already stated that that is a problem:
From a previous answer to you:
Because I don't think we have a sure methodology.
I think therefore I am, is only good for the conscious mind itself.
I can't prove that other people are conscious, although I'm 100% confident they are.
In exactly the same way we can't prove when we have a conscious AI.
But we may be able to prove that it is NOT conscious, which I think is clearly the case with current level AI. Although you don't accept the example I provided, I believe it is clear evidence of lack of a consciousness behind the high level of intelligence it clearly has.