So they don't actually have a definition of a AGI they just have a point at which they're going to announce it regardless of if it actually is AGI or not.
Great.
This is a most excellent place for technology news and articles.
So they don't actually have a definition of a AGI they just have a point at which they're going to announce it regardless of if it actually is AGI or not.
Great.
That's not a bad way of defining it, as far as totally objective definitions go. $100 billion is more than the current net income of all of Microsoft. It's reasonable to expect that an AI which can do that is better than a human being (in fact, better than 228,000 human beings) at everything which matters to Microsoft.
Good observation. Could it be that Microsoft lowers profits by including unnecessary investments like acquisitions?
So it'd take a 100M users to sign up for the $200/mo plan. All it'd take is for the US government to issue vouchers for video generators to encourage everyone to become a YouTuber instead of being unemployed.
If they actually achieve AGI I don't understand what money would even mean anymore. It essentially is just a mechanism for getting people to do things they don't otherwise want to do, if the AI can do it just as well as the human, but for free other than the electricity costs, why the hell would you pay a human to do it?
It's like saving up money, in case of nuclear war. There are a few particular moments in history where the state of the world on the far side of the event is so different to the world on this side of the event that there's no point making any kind of plans based on today systems.
I see what you're saying and I agree that if, for example, we get an AI god then money won't be useful. However, that's not the only possible near-future outcome and if the world as we know it doesn't end then money can be used by AIs to get other AIs to do something they don't otherwise want to do.
hence the worldcoin stuff - not just machine to machine. allows "ai" to perform real world action through human incentivization. entirely disturbing if you ask me.
My point is if AI takes over all of the work there won't be any jobs for humans. So they won't have any money.
So who are all the AI companies going to sell their products to? The whole system doesn't work in an AI future and we don't need AI gods to be able to do our jobs, after all most humans are idiots.
Also AI doesn't need motivation.
Trade (facilitated by money) doesn't require humans. It just requires multiple agents and positive-sum interactions. Imagine a company, run by an AI, which makes robots. It sells those robots to another company, also run by an AI, which mines metal (the robots do the mining). The robots are made from metal the first company buys from the second one. The first AI gets to make more robots than it otherwise would, the second AI gets to mine more metal than it otherwise would, and so both are better off.
They don't care that they're stuck in a loop, the same way humans keep creating new humans to create new humans to create new humans and so forth.
Does anyone have a real link to the non-stalkerware version of:
https://www.theinformation.com/articles/microsoft-and-openais-secret-agi-definition
-and the only place with the reference this article claims to cite but doesn't quote?