That might very well help, yes. However, two thoughts, neither at all well thought out:
If the Trump administration does fight OpenAI, let’s hope Altman doesn’t manage to judo flip the situation like he did with the OpenAI board saga, and somehow magically end up replacing Musk or Trump in the upcoming administration...
Musk’s own track record on AI x-risk is not great. I guess he did endorse California’s SB 1047, so that’s better than OpenAI’s current position. But he helped found OpenAI, and recently founded another AI company. There’s a scenario where we just trade extinction risk from Altman’s OpenAI for extinction risk from Musk’s xAI.
Potentially a hot take, but I feel like xAI’s contributions to race dynamics (at least thus far) have been relatively trivial. I am usually skeptical of the whole “I need to start an AI company to have a seat at the table”, but I do imagine that Elon owning an AI company strengthens his voice. And I think his AI-related comms have mostly been used to (a) raise awareness about AI risk, (b) raise concerns about OpenAI/Altman, and (c) endorse SB1047 [which he did even faster and less ambiguously than Anthropic].
The counterargument here is that maybe if xAI was in 1st place, Elon’s positions would shift. I find this plausible, but I also find it plausible that Musk (a) actually cares a lot about AI safety, (b) doesn’t trust the other players in the race, and (c) is more likely to use his influence to help policymakers understand AI risk than any of the other lab CEOs.
I’m sympathetic to Musk being genuinely worried about AI safety. My problem is that one of his first actions after learning about AI safety was to found OpenAI, and that hasn’t worked out very well. Not just due to Altman; even the “Open” part was a highly questionable goal. Hopefully Musk’s future actions in this area would have positive EV, but still.
I think that xAI’s contributions have been minimal so far, but that could shift. Apparently they have a very ambitious data center coming up, and are scaling up research efforts quickly. Seems very accelerate-y.
But he helped found OpenAI, and recently founded another AI company.
I think Elon’s strategy of “telling the world not to build AGI, and then going to start another AGI company himself” is much less dumb / ethical fraught, than people often credit.
I don’t think Vance is e/acc. He has said positive things about open source, but consider that thecontext was specifically about censorship and political bias in contemporary LLMs (bolding mine):
There are undoubtedly risks related to AI. One of the biggest:
A partisan group of crazy people use AI to infect every part of the information economy with left wing bias. Gemini can’t produce accurate history. ChatGPT promotes genocidal concepts.
The solution is open source
If Vinod really believes AI is as dangerous as a nuclear weapon, why does ChatGPT have such an insane political bias? If you wanted to promote bipartisan efforts to regulate for safety, it’s entirely counterproductive.
Any moderate or conservative who goes along with this obvious effort to entrench insane left-wing businesses is a useful idiot.
I’m not handing out favors to industrial-scale DEI bullshit because tech people are complaining about safety.
The words I’ve bolded indicate that Vance is at least peripherally aware that the “tech people [...] complaining about safety” are a different constituency than the “DEI bullshit” he deplores. If future developments or rhetorical innovations persuade him that extinction risk is a serious concern, it seems likely that he’d be on board with “bipartisan efforts to regulate for safety.”
That might very well help, yes. However, two thoughts, neither at all well thought out:
If the Trump administration does fight OpenAI, let’s hope Altman doesn’t manage to judo flip the situation like he did with the OpenAI board saga, and somehow magically end up replacing Musk or Trump in the upcoming administration...
Musk’s own track record on AI x-risk is not great. I guess he did endorse California’s SB 1047, so that’s better than OpenAI’s current position. But he helped found OpenAI, and recently founded another AI company. There’s a scenario where we just trade extinction risk from Altman’s OpenAI for extinction risk from Musk’s xAI.
Potentially a hot take, but I feel like xAI’s contributions to race dynamics (at least thus far) have been relatively trivial. I am usually skeptical of the whole “I need to start an AI company to have a seat at the table”, but I do imagine that Elon owning an AI company strengthens his voice. And I think his AI-related comms have mostly been used to (a) raise awareness about AI risk, (b) raise concerns about OpenAI/Altman, and (c) endorse SB1047 [which he did even faster and less ambiguously than Anthropic].
The counterargument here is that maybe if xAI was in 1st place, Elon’s positions would shift. I find this plausible, but I also find it plausible that Musk (a) actually cares a lot about AI safety, (b) doesn’t trust the other players in the race, and (c) is more likely to use his influence to help policymakers understand AI risk than any of the other lab CEOs.
I’m sympathetic to Musk being genuinely worried about AI safety. My problem is that one of his first actions after learning about AI safety was to found OpenAI, and that hasn’t worked out very well. Not just due to Altman; even the “Open” part was a highly questionable goal. Hopefully Musk’s future actions in this area would have positive EV, but still.
I think that xAI’s contributions have been minimal so far, but that could shift. Apparently they have a very ambitious data center coming up, and are scaling up research efforts quickly. Seems very accelerate-y.
I think Elon’s strategy of “telling the world not to build AGI, and then going to start another AGI company himself” is much less dumb / ethical fraught, than people often credit.
If Trump dies, Vance is in charge, and he’s previously espoused bland eaccism.
I keep thinking: Everything depends on whether Elon and JD can be friends.
I don’t think Vance is e/acc. He has said positive things about open source, but consider that the context was specifically about censorship and political bias in contemporary LLMs (bolding mine):
The words I’ve bolded indicate that Vance is at least peripherally aware that the “tech people [...] complaining about safety” are a different constituency than the “DEI bullshit” he deplores. If future developments or rhetorical innovations persuade him that extinction risk is a serious concern, it seems likely that he’d be on board with “bipartisan efforts to regulate for safety.”