Hacker Newsnew | past | comments | ask | show | jobs | submit | Miraste's commentslogin

If only there were some form of cheap, widely manufactured power generation technology that didn't use turbines... Are they really going to wait until 2030 to get more turbines rather than invest in solar?

I am clueless in this field, but solar seems to be unreliable and yield fraction of power required. Do you have a suggestion on something to read and learn more?

Reliability: complete solar deployment includes some form of power storage. There are many variations, but chemical battery technology is improving the fastest, so it's gaining the most ground.

Amount of power: World solar power generation capacity is in the terawatts and rapidly increasing, there's no issue with its potential ceiling. As a bonus, it tends to work best on land that's useless for other purposes.


Google china solar deployments to read about the logistics end of it

China's domestic chips are increasingly close to state-of-the-art. The US electrical grid is... not.

It's not at all clear whether automobiles were a net positive. They are more or less solely responsible for climate change (even emissions not directly from motor vehicles wouldn't be possible without them), which may prove to be the worst mistake in the history of technology.

What? 35B-A3B is not nearly as smart as 27B.

One interesting thing about Qwen3 is that looking at the benchmarks, the 35B-A3B models seem to be only a bit worse than the dense 27B ones. This is very different from Gemma 4, where the 26B-A4B model is much worse on several benchmarks (e.g. Codeforces, HLE) than 31B.

> This is very different from Gemma 4, where the 26B-A4B model is much worse on several benchmarks (e.g. Codeforces, HLE) than 31B.

Wouldn't you totally expect that, since 26A4B is lower on both total and active params? The more sensible comparison would pit Qwen 27B against Gemma 31B and Gemma 26A4B against Qwen 35A3B.


They're comparing Qwen's moe vs dense (smaller difference) against Gemma's moe vs dense (bigger difference). Your proposed alternative misses the point.

Gemma's dense is bigger than its moe's total parameters. You could totally expect the moe to do terribly by comparison.

yeah the 27B feels like something completely different. If you use it on long context tasks it performs WAY better than 35b-a3b

I've been telling analysts/investors for a long time that dense architectures aren't "worse" than sparse MoEs and to continue to anticipate the see-saw of releases on those two sub-architectures. Glad to continuously be vindicated on this one.

For those who don't believe me. Go take a look at the logprobs of a MoE model and a dense model and let me know if you can notice anything. Researchers sure did.


Dense is (much) worse in terms of training budget. At inference time, dense is somewhat more intelligent per bit of VRAM, but much slower, so for a given compute budget it's still usually worse in terms of intelligence-per-dollar even ignoring training cost. If you're willing to spend more you're typically better off training and running a larger sparse model rather than training and running a dense one.

Dense is nice for local model users because they only need to serve a single user and VRAM is expensive. For the people training and serving the models, though, dense is really tough to justify. You'll see small dense models released to capitalize on marketing hype from local model fans but that's about it. No one will ever train another big dense model: Llama 3.1 405B was the last of its kind.


You want to take bets on this? I'm willing to bet 500USD that an open access dense model of at least 300B is released by some lab within 3 years.

MoE isn't inherently better, but I do think it's still an under explored space. When your sparse model can do 5 runs on the same prompt in the same time as a dense model takes to generate one, there opens up all sorts of interesting possibilities.

Yes.

I can see analyzing it from a psychological perspective as a means of predicting its behavior as a useful tactic, but doing so because it may have "experiences or interests that matter morally" is either marketing, or the result of a deeply concerning culture of anthropomorphization and magical thinking.

An understandable reaction, but, qua philosopher, it brings me no joy to inform you that most of the things we did with a computer in 2020 are 'anthropomorphized', which is to say, skeumorphic, where the 'skeu' is human affect. That's it; that's the whole thing; that's what we're building.

To the extent that AI is a successful interface, it will necessarily be addressable in language previously only suited to people. So it is responsible to begin thinking of it as such, even tendentiously, so we don't miss some leverage that our wetware could see if we thought about it in that way.

Think of it as sort of like modelling a univariate function on a 2D Cartesian plane -- there is nothing 'in' the u-func that makes it graphable, but, by enabling us to recruit specialized optic-chiasm subsystems, it makes some functions much, much easier to reason about.

Similarly, if you can recruit the millions (billions?) of evolution-years that were focused on detecting dangerous antisocial personalities and tendencies, you just might spot something important in an AI.

It's worth doing for the precautionary principle alone, if not for the possibility of insight.


> a deeply concerning culture of anthropomorphization and magical thinking.

That’s the reverse Turing test. A human that can’t tell that it’s talking to a machine.


They are a for-profit company, working on a project to eliminate all human labor and take the gains for themselves, with no plan to allow for the survival of anyone who works for a living. They're definitionally not your friends. While they remain for-profit, their specific behaviors don't really matter.

I work for a tech company that eliminates a form of human labour and they remain for profit

Sure, most tech companies eliminate some form of human labor. Anthropic aims to eliminate all human labor, which is very different.

>We plan to launch new safeguards with an upcoming Claude Opus model, allowing us to improve and refine them with a model that does not pose the same level of risk as Mythos Preview2.

This seems like the real news. Are they saying they're going to release an intentionally degraded model as the next Opus? Big opportunity for the other labs, if that's true.


The other labs already censor their models. Everyone is trying to find the sweet spot where performance and ‘alignment’ are both maximized. This seems no different

> Big opportunity for the other labs, if that's true.

It sounds like this is considered military grade technology as cryptography in the 90s. The big difference is it's very expensive to create, and run those models. It's not about the algorithm. If the story rhymes it could be a big opportunity to other regions in the world.


Well since Anthropic treats us as second class evil citizens, I guess they don't want our evil money either.

Child car seat regulations are state laws passed by state politicians. They are not experts in any sense of the word, and generally don't bother with evidence or studies when creating said laws.


Just follow the CDC recommendation then, which is to keep kids in a car seat until 11-12 years old.

I'm not arguing all laws are good or make sense. In this specific case, the law lines of up with the recommendation of experts studying the topic.


AI is much better at generating text that resembles scientific papers than it is at literary writing. Even if they're not all flagged as AI, the incidence will be much lower because they're simply bad writing. They won't make it out of the slush pile at places like GP listed.


There is a line between a poker game with friends, or even a professional poker industry, and a sophisticated tech company operating a nationwide low-friction gambling app, incentivized to optimize harming its users as much as possible. This line was enshrined into law until recently.


I agree, which is why I think it is going too far to say “anyone who works in the gambling industry is bad”


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: