Hacker Newsnew | past | comments | ask | show | jobs | submit | bertili's commentslogin

Exactly. The emperor has no clothes. The largest investments in US tech in history and yet there less than a year of moat. OpenAI or Anthropic will not be able to compete with Chinese server farms and so the US strategy is misplaced investments that will come home to roast.

And we will have Deepseek 4 in a few days...


Surely this is the elephant in the room, but the point here is that Apple as control over its ecosystem, so it may be able to sandbox and make entitlements and transparency good enough, in the apps that the bot can access.

Like I said: sandboxing doesn't solve the problem.

As long as the agent creates more than just text, it can leak data. If it can access the internet in any manner, it can leak data.

The models are extremely creative and good at figuring out stuff, even circumventing safety measures that are not fully air tight. Most of the time they catch the deception, but in some very well crafted exploits they don't.


The other realistic setup is $20k, for a small company that needs a private AI for coding or other internal agentic use with two Mac Studios connected over thunderbolt 5 RMDA.


That won’t realistically work for this model. Even with only ~32B active params, a 1T-scale MoE still needs the full expert set available for fast routing, which means hundreds of GB to TBs of weights resident. Mac Studios don’t share unified memory across machines, Thunderbolt isn’t remotely comparable to NVLink for expert exchange, and bandwidth becomes the bottleneck immediately. You could maybe load fragments experimentally, but inference would be impractically slow and brittle. It’s a very different class of workload than private coding models.


People are running the previous Kimi K2 on 2 Mac Studios at 21tokens/s or 4 Macs at 30tokens/s. Its still premature, but not a completely crazy proposition for the near future, giving the rate of progress.


> 2 Mac Studios at 21tokens/s or 4 Macs at 30tokens/s

Keep in mind that most people posting speed benchmarks try them with basically 0 context. Those speeds will not hold at 32/64/128k context length.


If "fast" routing is per-token, the experts can just reside on SSD's. the performance is good enough these days. You don't need to globally share unified memory across the nodes, you'd just run distributed inference.

Anyway, in the future your local model setups will just be downloading experts on the fly from experts-exchange. That site will become as important to AI as downloadmoreram.com.


Depends on if you are using tensor parallelism or pipeline parallelism, in the second case you don't need any sharing.


RDMA over Thunderbolt is a thing now.


I'd love to see the prompt processing speed difference between 16× H100 and 2× Mac Studio.


Prompt processing/prefill can even get some speedup from local NPU use most likely: when you're ultimately limited by thermal/power limit throttling, having more efficient compute available means more headroom.


I asked GPT for a rough estimate to benchmark prompt prefill on an 8,192 token input. • 16× H100: 8,192 / (20k to 80k tokens/sec) ≈ 0.10 to 0.41s • 2× Mac Studio (M3 Max): 8,192 / (150 to 700 tokens/sec) ≈ 12 to 55s

These are order-of-magnitude numbers, but the takeaway is that multi H100 boxes are plausibly ~100× faster than workstation Macs for this class of model, especially for long-context prefill.


You do realize that's entirely made up, right?

Could be true, could be fake - the only thing we can be sure of is that it's made up with no basis in reality.

This is not how you use llms effectively, that's how you give everyone that's using them a bad name from association


That's great for affordable local use but it'll be slow: even with the proper multi-node inference setup, the thunderbolt link will be a comparative bottleneck.


The "Deepseek moment" is just one year ago today!

Coincidence or not, let's just marvel for a second over this amount of magic/technology that's being given away for free... and how liberating and different this is than OpenAI and others that were closed to "protect us all".


There's been so many moments that folks not really heavy into LLM have missed, DeepSeekR1 was great, but so was all the "incremental" improvements, v3-0324, v3.1, v3.1-terminus, and now v3.2-speciale. With that this is the 3rd great Kimi model, then GLM has been awesome, since 4.5, with 4.5, 4.5-air, 4.6, 4.7 and now 4.7 flash. Minimax-M2 has also been making waves lately. ... and i'm just talking about the Chinese model without adding the 10+ Qwen models. Outside of Chinese models, mistral-small/devstral, gemma-27b-it, gpt-oss-120b, seed-os have been great, and I'm still talking about just LLM, not image, audio or special domain models like deepseek-prover and deepseek-math. It's really a marvel what we have at home. I cancelled OpenAI and Anthropic subscription 2 years ago once they started calling for regulation of open models and I haven't missed them one bit.


What's your hardware/software setup?

It’s not coincidence. Chinese companies tend to do big releases before Chinese new year. So expect more to come before Feb 17.


What amazes me is why would someone spend millions to train this model and give it away for free. What is the business here?


Chinese state that maybe sees open collaboration as the way to nullify any US lead in the field, concurrently if the next "search-winner" is built upon their model the Chinese worldview that Taiwan belongs to China and Tiamen Square massacre never happened.

Also their license says that if you have a big product you need to promote them, remember how Google "gave away" site searche widgets and that was perhaps one of the major ways they gained recognition for being the search leader.

OpenAI/NVidia is the Pets.com/Sun of our generation, insane valuations, stupid spend, expensive options, expensive hardware and so on.

Sun hardware bought for 50k USD to run websites in 2000 are less capable than perhaps 5 dollar/month VPS's today?

"Scaling to AGI/ASI" was always a fools errand, best case OpenAI should've squirreled away money to have a solid engineering department that could focus on algorithmic innovations but considering that Antrophic, Google and Chinese firms have caught up or surpassed them it seems they didn't.

Once things blows up, those closed options that had somewhat sane/solid model research that handles things better will be left and a ton of new competitors running modern/cheaper hardware and just using models are building blocks.


> "Scaling to AGI/ASI" was always a fools errand

Scaling depends on hardware, so cheaper hardware on a compute-per-watt basis only makes scaling easier. There is no clear definition of AGI/ASI but AI has already scaled to be quite useful.


> There is no clear definition of AGI

? Yes there's a clear defintion

https://en.wikipedia.org/wiki/Artificial_general_intelligenc...

I think you mean there is no agreed upon definition? Even that is dubious statement.


> Taiwan belongs to China

So they are on the same page as the UN and US?

The One China policy refers to a United States policy of strategic ambiguity regarding Taiwan.[1] In a 1972 joint communiqué with the PRC, the United States "acknowledges that all Chinese on either side of the Taiwan Strait maintain there is but one China and that Taiwan is a part of China" and "does not challenge that position."

https://en.wikipedia.org/wiki/One_China https://en.wikipedia.org/wiki/Taiwan_and_the_United_Nations


The One China policy is a fiction of foreign policy statecraft, designed to sideline the issue without having to actually deal with it. It is quite clear that apart from the official fiction there is a real policy that is not One China. This is made clear by the weapons sales to Taiwan that specifically calibrated to make a Chinese military action harder.


If it's a fiction https://www.congress.gov/crs-product/IF12503 then it's a very deceitful and and elaborate one.

Existence of an independent and effectively sovereign state on the island of Taiwan (however one calls it) is a fact. Whatever doublespeak governments of other countries or international organizations engage in due to political reasons does not change that.


No, you are misinformed. The facts are that both the Chinese mainland authority (the CPC) and the government of Chinese Taipei each lay claim to the entirety of the mainland+island landmass. Get that? They both regard themselves and themselves alone as being sovereign over the unified territory.

That is why Chinese Taipei is not independent and why Beijing has said that if they threaten independence (that is to say, that if Chinese Taipei relinquishes its claim over the mainland -- that it in a way secedes from the union state) that the mainland would move to take full political control over the island.

It's a highly unusual situation -- there are countries which have governments in exile, this is not that. You can think of this like one country with two separate competing governments which irreconcilable claims. Clearly given the mainland's economic (and consequent military rise) Chinese Taipei cannot take over the mainland and again considering the power differential they would be insane to declare full independence. Why is why this bizarre situation persists.

Furthermore third-party countries which recognise Chinese Taipei's independence is shrinking and shrinking (and so it should be) because properly speaking the sane thing to do is recognise that there is only one China (because both the the authority of mainland China (the CPC) and the government of Taiwan assert that there is but one China) but deal with each authority separately. Economically speaking and politically speaking this is how things play out.

Because the situation is so unusual I think it is fair to say that most people just assume Taiwan is a breakaway province. And it's not surprising that they do because that's what happens normally in these situations. That would be the case if Taiwan amended it's constitution to explicitly relinquish its claim over the mainland but they are not going to do that unless they are prodded to by an outside third-party or if they suddenly get a wave of nationalism or if they suddenly become collectively suicidal. So this highly unusual and bizarre status quo persists.

This is my interaction with ChatGPT (obviously it would be pointless talking to DeepSeek, etc about this)

===

ME: does chinese taipei lay claim to mainland china?

ChatGPT: Yes — but it depends on *which “Chinese Taipei” you mean*:

   ###  \*Chinese Taipei as the Republic of China (Taiwan)\*

   \* The government based in Taiwan is officially the \*Republic of China (ROC)\*. Under its \*old constitution\* (dating back to when the ROC ruled all of China), it *theoretically* still contains language implying sovereignty over all of China — including the mainland. In practice, modern Taiwanese governments (especially under the Democratic Progressive Party) do \*not actively assert a real claim to govern mainland China\* and focus on governing Taiwan and nearby islands. ([House of Commons Library][1])
* So while there is *historical constitutional language* about the ROC’s territorial scope that could be interpreted to include mainland China, *Taipei today does not pursue actual governance or enforcement of such a claim* in international relations.

   ###  \*People’s Republic of China (PRC) claims Taiwan\*

   \* The \*PRC\* (mainland China) claims that Taiwan is part of China and should eventually be “reunified,” insisting on a \*One-China principle\* that includes Taiwan under its sovereignty. This is the basis of most countries’ diplomatic relations with Beijing instead of Taipei. ([House of Commons Library][1])

   ###  Key nuance

   \* Taiwan (the ROC government) effectively functions as a \*separate political entity with its own government, military, and democratic system\* and does \*not control or claim to govern mainland China’s territory in practice\* even if older constitutional wording could be interpreted that way. ([House of Commons Library][1])
* Meanwhile, the PRC claims Taiwan as part of its territory, a position Taiwan strongly disputes. ([taipeitimes.com][2])

[1]: https://commonslibrary.parliament.uk/research-briefings/cdp-... "International status of Taiwan - House of Commons Library"

[2]: https://www.taipeitimes.com/News/front/archives/2025/03/08/2... "Taiwan rejects China’s UN claim - Taipei Times"

ME: What does "it theoretically still contains language implying" mean? It either does or it doesn't.

ChatGPT: You’re right to push back — that phrasing was imprecise. Let’s be exact.

   ### Short answer

   \*Yes, the Republic of China (ROC) constitution *does* still claim all of China, including the mainland.\*
That is a matter of constitutional fact, not theory.

   ### What the constitution actually says

   \* The \*ROC Constitution (1947)\* was written when the ROC claimed to be the sole government of *all* China.
* It *defines the state’s territory as China*, without listing boundaries or excluding the mainland. * *There has never been a constitutional amendment redefining the ROC’s territory as limited to Taiwan, Penghu, Kinmen, and Matsu.*

   So legally speaking:

   > \*Mainland China remains part of ROC territory under the constitution.\*

   ### Where the confusion comes from

   The confusion arises because of \*later constitutional add-ons\*, not because the claim disappeared.

   Since the 1990s, Taiwan adopted the \*Additional Articles of the Constitution\*, which:

   \* Suspend application of many provisions *with respect to the mainland*
* Define elections, offices, and governance as applying only to the “*free area of the Republic of China*”

   Crucially:

   \* These articles \*do not renounce the mainland\*
* They merely acknowledge that the ROC *does not currently govern it*

   So the situation is:

   | Aspect                             | Status                                |
   | ---------------------------------- | ------------------------------------- |
   | Constitutional territorial claim   | \*All of China (including mainland)\* |
   | Practical governance               | \*Taiwan and nearby islands only\*    |
   | Active effort to reclaim mainland  | \*No\*                                |
   | Formal constitutional renunciation | \*No\*                                  |

   ### Why people hedge with language

   People (especially diplomats and scholars) often say things like *“theoretically implies”* to avoid the blunt conclusion that:

   \* \*Taiwan’s constitution still claims mainland China\*, which is politically awkward given Taiwan’s current identity and international position.

   But legally, there’s no ambiguity:

   \* \*It either does or it doesn’t — and it does.\*

   ### One-sentence summary

   \*Chinese Taipei (the ROC) constitutionally claims mainland China, but politically and practically does not pursue or assert that claim.\*

   If you want, I can also explain \*why Taiwan has never amended the constitution to drop the claim\*, because that’s a fascinating mix of law, identity, and geopolitics.
===

(Sorry for the dreadful formatting). I am more than happy to be corrected on this.

This dark side of all this is that the original inhabitants of the island are not "Chinese" (Han) and the island has had both Japanese and Chinese rule if I'm not mistaken: https://en.wikipedia.org/wiki/Taiwanese_indigenous_peoples


None of this changes the fact that there is an independent country on the island of Taiwan. It really doesn't matter who you or someone else is calling it, that's just an observable fact.

I love how Tiananmen square is always brought up as some unique and tragic example of disinformation that could never occur in the west, as though western governments don't do the exact same thing with our worldview. Your veneer of cynicism scarcely hides the structure of naivety behind.


The difference is that, in the west, there's an acceptable counter narrative. I can tell you that Ruby Ridge and Waco never should've happened and were examples of government overreach and massacre of it's own citizens. Or <insert pet issue with the government here>

You can't with Tiananmen square in China


I still see/hear cynicism with a hidden structure of naivety behind.


> Tiananmen square

The Tiananmen Square killings happened independent of one's "worldview".

What's an example of "the exact same thing" done to OpenAI models by western governments?


Speculating: there are two connected businesses here, creating the models, and serving the models. Outside of a few moneyed outliers, no one is going to run this at home. So at worst opening this model allows mid-sized competitors to serve it to customers from their own infra -- which helps Kimi gain mindshare, particularly against the large incumbents who are definitely not going to be serving Kimi and so don't benefit from its openness.

Given the shallowness of moats in the LLM market, optimizing for mindshare would not be the worst move.


Moonshot’s (Kimi’s owner) investors are Alibaba/Tencent et al. Chinese market is stupidly competitive, and there’s a general attitude of “household name will take it all”. However getting there requires having a WeChat-esque user base, through one way or another. If it’s paid, there’ll be friction and it won’t work. Plus, it undermines a lot of other companies, which is a win for a lot of people.


I think this fits into some "Commoditize The Complement" strategy.

https://gwern.net/complement


I think there is a book (Chip War) about how the USSR did not effectively participate in staying at the edge of the semiconductor revolution. And they have suffered for it.

China has decided they are going to participate in the LLM/AGI/etc revolution at any cost. So it is a sunk cost, and the models are just an end product and any revenue is validation and great, but not essential. The cheaper price points keep their models used and relevant. It challenges the other (US, EU) models to innovate and keep ahead to justify their higher valuations (both monthly plan, and investor). Once those advances are made, it can be bought back to their own models. In effect, the currently leading models are running from a second place candidate who never gets tired and eventually does what they do at a lower price point.


In some way, the US won the cold war by spending so much on military that the USSR, in trying to keep up, collapsed. I don't see any parallels between that and China providing infinite free compute to their AI labs, why do you ask?


All economically transformative technologies have done similar. If it's privatized, it's not gonna be transformative across the industry. The GPS, the internet, touchscreens, AI voice assistants, microchips, LCDs, etc were all publicly funded (or made by Bell Labs which had a state-mandated monopoly that forced them to open up their patents).

The economist Mariana Mazzucato wrote a great book about this called The Entrepreneurial State: Debunking Public vs. Private Sector Myths


> What amazes me is why would someone spend millions to train this model and give it away for free. What is the business here?

How many millions did Google spend on Android (acquisition and salaries), only to give it away for free?

Usually, companies do this to break into a monopolized market (or one that's at risk of becoming one), with openness as a sweetener. IBM with Linux to break UNIX-on-big-iron domination, Google with Android vs. iPhone, Sun with OpenSolaris vs. Linux-on-x86.


Hosting the model is cheaper per token, the more batched token you get. So they have big advantage here.


Curious to hear what “OpenAI” thinks the answer to this is


It's another state project funded at the discretion of the party.

If you look at past state projects, profitability wasn't really considered much. They are notorious for a "Money hose until a diamond is found in the mountains of waste"


I am convinced that was mostly just marketing. No one uses deepseek as far as I can tell. People are not running it locally. People choose GPT/Gemini/Claude/Grok if you are giving your data away anyway.

My biggest source of my conspiracy is that I made a reddit thread asking a question: "Why all the deepseek hype" or something like that. And to this day, I get odd, 'pro deepseek' comments from accounts only used every few months. Its not like this was some highly upvoted topic that is in the 'Top'.

I'd put that deepseek marketing on-par with an Apple marketing campaign.


I don't use DeepSeek, but I prefer Kimi and GLM to closed models for most of my work.


Except that, In OpenRouter, Deepseek always maintain in Top 10 Ranking. Although I did not use it personally, i believe that their main advantage over other model is price/performance.


Fifth in market share in fact!

https://openrouter.ai/rankings

There are a lot of applications where you really just want a cheap and efficient model that's still somewhat competitive and that's exactly the niche DeepSeek fulfills the best.


I mean, there are credible safety issues here. A Kimi fine-tune will absolutely be able to help people do cybersecurity related attacks - very good ones.

In a few years, or less, biological attacks and other sorts of attacks will be plausible with the help of these agents.

Chinese companies aren't humanitarian endeavors.


What a time to be (still) alive.

When we look back in a few years and ask the question: who actually got to pay for the Epstein crimes and coverups, we come to the surprising answer it is the Greenlandes and other innocent societies that got ripped apart by this maniac and his supporters.


Also the entire economy. Stocks are at an all time high on both sides of the Atlantic but the real world economy is struggling.


It's not a failure of IP6 but a failure of society.

We all thought the internet would become decentralized and that everyone should have an IP and a funky website. But instead social media took over, big tech and a few big discussion sites where we all must fit in a digital life and watch ads and share our data to become a good product for all the others to consume.


"The missile is too round at the top, it needs to be pointy. Round missiles are not scary"

- The Dictator



Thanks! Macroexpanded:

SHARP, an approach to photorealistic view synthesis from a single image - https://news.ycombinator.com/item?id=46284658 - Dec 2025 (108 comments)


I love how virtually no GitHub instructions related to AI simply work as written.

Each assumes you already have their developer environment configured to have the tool work, but simply don’t have it compiled yet.


This repo's instructions seem to work as written just fine?


Unfortunately, this will not happen. Even if they have it all working:

Above all, Apple wants to show that their hardware is awesome, especially because it really is. Running x86 games or compatibility layers even with great emulation will make that $3000 Mac look half decent at best, against a $1500 gaming laptop. Simply not the story Apple want to tell.


If apple wanted to show that they have good hardware they wouldn't gimp the iPad pro with iOS. They really don't care.


Currently, someone interested in an iPad and needing the power of a MB, will have to buy both.

If they stopped restricting the iPad, those people would only have to buy an iPad.

And as someone without a single interest in an iPad, I would worry that removing the iPad limitations would increase its market-share and lead to Apple reducing even more their interest in the MB, which would be terrible news to me.


I used to understand/agree with this point, but over the past few years i've transitioned to my ipad pro for mobile usage and it has become my daily driver for mobile computing. When i need macos for anything, i typically will use Jump to connect and do something real quick, but that's rare. I'm starting to not understand why i wouldn't just want an ipad pro running a touch friendly (and i mean it would have to be VERY touch friendly) version of macos. again, i would have normally agreed with you, but that line is starting to blur for me...


Two 512GB Mac Studios connected with thunderbolt 5.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: