It's rather obviously illegal to leak classified intel by taking public actions based off of it... that's practically the meaning of the word "classified".
I for one delight in bicycles where neither wheel can turn!
It continues to amaze me that these models that definitely know what bicycle geometry actually looks like somewhere in their weights produces such implausibly bad geometry.
Also mildly interesting, and generally consistent with my experience with LLMs, that it produced the same obvious geometry issue both times.
> It continues to amaze me that these models that definitely know what bicycle geometry actually looks like somewhere in their weights produces such implausibly bad geometry.
I feel like the main problem for the models is that they can't actually look at the visual output produced by their SVG and iterate. I'm almost willing to bet that if they could, they'd absolutely nail it at this point.
Imagine designing an SVG yourself without being able to ever look outside the XML editor!
> Imagine designing an SVG yourself without being able to ever look outside the XML editor!
I honestly think I could do much better on the bicycle without looking at the output (with some assistance for SVG syntax which I definitely don't know), just as someone who rides them and generally knows what the parts are.
Performing the charity work of discovering bugs before someone evil uses them to cause damage does not somehow obligate you to perform more charity and fix those bugs.
Is this cash or compute? Elon has one of the world's biggest compute clusters spun up, and little inference demand to speak of.
Trading billions worth of idle compute, in exchange for a high-strike call option on the #3 player in the most-promising-vertical for AI, plus (presmuably) some access to their data, starts to sound like not a bad trade. Especially if you're pre-committed to betting your entire rocket company on winning in AI, and you're currently in sixth or seventh place.
It's true he could write off xAI today and the company could still fetch a trillion-dollar valuation. But I was more referring to his stated intentions - between his stated plans, his actions taking SpaceX from a profitable company to spending basically all their revenue (plus a rumored large chunk of what's raised via its IPO) on AI, and seeing his tendency to make bet-the-farm bets on Tesla, I think it's fair to say he's committing to bet all of SpaceX on xAI.
I heard he made a deal with a company to use his clusters. Is there good data on demand for Grok? Seems like relatively little chatter at least, in spite of tremendous investment.
He had a very close, decades long friendship with the most notorious sex-trafficker-of-children-to-rich-creeps in modern history for decades. And when imprisoned, that infamous pedophile died while in a federal prison under Trump's control, with a strange gap in the CCTV video footage. And Trump's handling of the entire Epstein Files saga makes it clear that Trump is described extensively in those files and he desperately wants to conceal it. What could be in there that he would use the entire justice department to try and redact? Trump is shameless about things that are legal even if they're salacious (like sleeping with porn star Stormy Daniels), so you have to wonder, what could Jeffery Epstein's good friend be trying to conceal?
Also, he owned the Miss Universe org (including Miss USA and Miss Teen USA) for decades, and he was known to walk into the dressing rooms of teen contestants as young as 15 while they were undressed. [0]
Also, he bragged about molesting women, and a court of law found that he sexually assaulted E Jean Carroll.
I haven't proven the case that Trump had sex with a minor, but there's way more than enough probable cause to believe it's more likely than not.
Imagine there's a camera continuously recording a cookie jar. A child eats all of the cookies and then deletes the footage from the time they ate the cookies. A parent returns to find their child covered in crumbs, loudly proclaiming they haven't eaten a cookie in years and actively interferes with the parent's investigation and tries to distract from it by throwing a brick through the window of an Iranian family down the street.
Are any of the facts in this hypothetical "evidence"? With the knowledge of the truth (that the kid ate the cookies), it's clear these are all relevant pieces of evidence. If we take knowledge of the truth out of the equation, would these facts still be evidence? Unambiguously they would.
Definitionally both circumstantial and direct evidence are forms of evidence. No modifier is necessary.
And incidentally you can be convicted in a court of law purely on circumstantial evidence, and that's the place in society where we have the highest standard of proof. The evidence all being circumstantial is not a gotcha.
This isn't court. The evidence, such as it is, is all of the smoke which commonly motivates people to look for fire. The strongest and most comprehensive that I've seen is the argument that if Trump was not implicated in the Epstein files, he would be publishing them in free book form himself and forcing every media outlet to advertise it. Slight exaggeration, but I think truly only slight.
Not really relevant to the thread, but there are simple answers to the "eViDeNcE??" question. You may have already known this.
Has the availability of deepfake porn generation reduced the demand for deepfake porn featuring real people? When deepfake generators are capable of creating convincing imagery of flawless ideal fake humans, why do you suppose there’s so many real humans who report being non-consensual subjects of deepfake porn?
> Has the availability of deepfake porn generation reduced the demand for deepfake porn featuring real people?
yes
> When deepfake generators are capable of creating convincing imagery of flawless ideal fake humans, why do you suppose there’s so many real humans who report being non-consensual subjects of deepfake porn?
> Doesn't have to be. You can train it on normal pictures of children and nude images of adults.
You say this so casually, as though it were a normal thing to know, or as if a normal person would know it. Does that actually seem true where you live right now?
And how do you know that, anyway, Harsh? I mean, all those "unblocked" games you stole to give away and that you also put on Github, that's one thing. But this...
Come on, it's not hard to come up with this idea. And it's not even true, model trained on clothed children and nude adults wouldn't know how children's genitals look like.
Problem is basically, that if the option works out (Cursor truly has the talent to train a frontier model on SpaceX's infrastructure, and were simply lacking the infra before) the fair price would be way way more than $60B.
OpenAI tried to acquire Windsurf last year for $3B and couldn't.
1) A gamble based on cursor's compute constraint
2) if 1) plays out, he can purchase cursor via shares of spaceX over valued shares, at a fixed price should the valuation increase.
This is a bit simplistic. It's the VS Code that everyone used before cc came to town. Real devs, on real projects. All that data they collected is worth a lot more than "just vscode". Their composer2 is better than kimi2.5 and it's just a finetune on that data.
xAI had a decent model in grok4 (it was even sota on a bunch of benchmarks for a few weeks), but they didn't have great coding models (code-fast was ok-ish but nothing to write home about, certainly nowhere near SotA). Now that they've been banned from using claude, they'll get their expertise + data to build a coding model on top of whatever grok5 will be + their cluster for compute.
It doesn't sound like a bad plan to me, financial shenanigans or not.
There's a lengthy discussion to be had here, and there's enough lawyerspeak in every provider's data retention policy to wiggle out of anything. A few notes from their current data use page:
> If you enable “Privacy Mode” in Cursor’s settings: zero data retention will be enabled for our model providers. Cursor may store some code data to provide extra features. None of your code will ever be trained on by us or any third-party.
Note the "may store some code data" and "none of your code will ever be trained on". In general you never want to include actual customer code in training the data, because of leaks that you may not want. Say someone has a hash somewhere, and your model autocompletes that hash. Bad. But that's not to say you couldn't train a reward model on pairs of prompts + completions. You have "some code data" (which could be acceptance rate) and use that. You just need to store the acceptance rate. And later, when you train new models, you check against that reward model. Does my new model reply close enough to score higher? If so, you're going in the right direction.
> If you choose to turn off “Privacy Mode”: we may use and store codebase data, prompts, editor actions, code snippets, and other code data and actions to improve our AI features and train our models.
Self explainatory.
> Even if you use your API key, your requests will still go through our backend!
They are collecting data even if you BYOK.
> If you choose to index your codebase, Cursor will upload your codebase in small chunks to our server to compute embeddings, but all plaintext code for computing embeddings ceases to exist after the life of the request. The embeddings and metadata about your codebase (hashes, file names) may be stored in our database.
They don't store (nor need to store) plain text, but they may store embeddings and metadata. Again, you can use those to train other things, not necessarily models. You can use metadata to check if you're going in the right direction.
Not quite first party, but composer 2 is far superior to grok for coding. Unless you're eluding to them using SpaceX infra to train their own model vs. using grok
Not only is it almost certainly compute (“services”) it’s likely priced at Anthropic rack-rate, or at least what Cursor’s been paying Anthropic.
The cluster’s already paid for, so likely in the $2B range for operating cash needs. Not more than $5.
If I imagine bringing in Cursor’s team to build a frontier model, ideally combined with Grok, which has one of the few truly proprietary data feeds available to it, and with a much larger custom model Cursor can solidify a place, and I get to do a stock swap to buy it, this sounds like a bet worth making.
Upshot - I bet there’s an MS/oAI deal on IP on the back of this; meanwhile the cluster goes brrr.
I'm not sure what you're referring to by "that" but I think you're right that it's 10B to not purchase or 60B to purchase, so as an option posting $10B for an option with a $50 strike price.
But they also get a whole bunch of AI Services from Cursor. Other comments have noted that xAI has fallen on bad times (idk one way or the other) so perhaps they were going to spend $5B on getting these services elsewhere, anyway.
SpaceX spending $1B a month on various AI services seems ~plausible
(EDIT - Or maybe it's an IP transfer, or maybe it's over a longer time horizon. Idk but SpaceX clearly expects value from 'our work together' even if they don't exercise.)
$1B per month on AI services does not seem remotely plausible to me... Engineers don't consume that many tokens...
And on the AI development side they're the ones providing compute in the form of a "million H100 equivalent Colossus training supercomputer"... On top of the cash.
Cursor has no AI services, they do not develop their own frontier models. I see no reason to understand why $10bn for Cursor's services is an advantage xAI versus say a $10bn deal with Anthropic, OpenAI or Google.
It's true that Cursor doesn't have their own frontier models, but they are training their own models. They just aren't at frontier level yet. The $60B/$10B deal looks like a bet that this is a capital/GPU constraint rather than a capability one.
For a truly malicious actor, you're right. But it shifts it from "well we aren't obviously committing fraud by quantizing this model and not telling people" to "we're deliberately committing fraud by verifying our deployment with one model and then serving customer requests with another".
I suspect there's a lot of semi-malicious actors who are only happy to do the former.
> And if that’s the case, would it happen much quicker since technology has been speeding up year by year?
I wouldn't expect this.
Historically we've had a roughly exponential rate of shrinkage. If we keep that same exponential going, we should expect the amount of time to shrink "room full of compute" to "pocket full of compute" to be equal.
And recently we've fallen behind that exponential rate of shrinkage. And this is rather expected because exponentials are basically never sustainable rates of growth.
I still expect that technological progress is getting faster year by year, and that we're still shrinking compute, but that's not necessarily enough for the next shrinking to take less time than when we had exponential progress on shrinking.
Considering the propaganda value in controlling the inputs to the machine that answers peoples questions, I rather expect them to be subsidized forever.
Consider the propaganda value of a centrally-controlled apparatus like the iPhone, and then reflect on the 100%+ profit margins that product has enjoyed for the past decade.
reply