Naturally the guardrails cannot cover absolutely every possible specific use case, but they can cover most of the known potentially harmful scenarios under the normal, most common circumstances. If the companies won't do it themselves, then legislation can push them to do it, for example making them liable, if their LLM does something harmful. Regulating AI is not anti-AI.
Benedict_Espinosa
Probably the same kind of guardrails that they already have - teaching LLMs to recognise patterns of potentially harmful behaviour. There's nothing impossible in that. Shutting LLMs down altogether is a straw man and extreme example fallacy, when the discussion is about regulation and guardrails.
Discussing the damage LLMs do does not, of course, in any way negate the damage that social media does. These are two different conversations. In the case of social media there's probably government regulation needed, as it's clear by now that the companies won't regulate themselves.
It's not about banning or refusing AI tools, it's about making them as safe as possible and regulating their usage.
Your argument is the equivalent of "guns don't kill people" or blaming drivers for Tesla's so-called "full self-driving" errors leading to accidents, because "full self-driving" switches itself off right before the accident, leaving the driver responsible as the one who should have paid more attention, even if there was no time left for him to react.
It's their money to invest however they want only if it comes with no strings attached, no obligations to use it for a specific purpose. We don't know if this is the case, so there's no basis to argue that they can do whatever they want with it.
What is corruption? It's a form of dishonesty that is undertaken by a person or an organization that is entrusted in a position of authority. That certainly seems to be the case here - not by SpaceX and xAi as such, but by Musk and his involvement in the government.
And Musk has everything to do with Grok.
A kind of computerised profascist authoritarian dystopia, a combination "1984" and "Brave New World" with technocratic oligarchy, total surveillance, killer robots and unsafe self-driving cars in the world increasingly subject to natural catastrophes due to steadily worsening climate change.
Name a thing that is unbiased. It's generally significantly less biased than humans are.
What could they realistically do, when Trump controls all branches of government from Congress to the Supreme Court? It can be argued that they should have shut down his government in March, when they had the chance to reject the spending bill - but what can they do now?
It's not generally a worse chatbot than others - that is, until Musk manages to ruin it completely.
Hence the updates. This is something that Musk has been trying to overturn for some time now.
They love him exactly for what he is in their eyes: strong, candid and clever. For them he is honest ("he tells it like it is") and can do no wrong ("Trump was right about everything"). What other people see as idiocy, weakness and corruption, they see as strengths.
Big part of this appeal is that his persona gives his followers an unlimited license to be as dumb, fascist and hateful as they want.
European politics is far from perfect, although it is arguably better and less corrupt than in the US now. But it was not politics that killed European mobile phone industry - it was competition along with mismanagement and miscalculations of the European mobile phone manufacturers. Symbian was just a weak and clumsy platform compared to iOS and Android, it could not compete in a changing market.
I do indeed, and I think that it's a remarkably disingenuous and biased take.