Hacker Newsnew | past | comments | ask | show | jobs | submit | woadwarrior01's commentslogin

HF is indeed banned in China. The Chinese equivalent of HF is ModelScope[1].

[1]: https://modelscope.cn/


There's also the recently released zvec[1], the tagline for which is: The SQLite of Vector Databases.

[1]: https://github.com/alibaba/zvec


Judging by the code in the HF transformers repo[1], smaller dense versions of this model will most likely be released at some point. Hopefully, soon.

[1]: https://github.com/huggingface/transformers/tree/main/src/tr...


Let's not forget the KV-cache which needs a lot of RAM too (although not as much as the model weights), and scales up linearly with sequence length.

I looked it up. It's called APX (Advanced Performance Extensions)[1].

[1]: https://www.intel.com/content/www/us/en/developer/articles/t...


Oh my, it has three-operand instructions now. VAX vindicated...

There's also Normalized Google Distance (a distance metric using the number of search results as a proxy), which can be used for text classification.

https://en.wikipedia.org/wiki/Normalized_Google_distance


My advisor in grad school had me implement a "typo distance" metric on strings once (how many single-key displacements for a typist using home row touch-typing to get from string A to string B), which seemed kind of cool. I never did find out what if anything she wanted to use it for.

Probably for spell checking! If you encounter an unknown word, you can rank candidate replacements by your advisor's metric.

100%. Reddit and X are surreptitiously the real Moltbooks. :)

I remember the time when Python was the underdog and most of AI/ML code was written in the Matlab or Lua (torch). People would roll their eyes when you told them that you were doing deep learning with Python (theano).

What community hacks?


What I meant is, if you can somehow get it working, it is not currently a supported first-party thing, not that I am aware of such thing existing.


That's a very tenuous analogy. Microcontrollers are circuits that are designed. LLMs are circuits that learned using vast amounts of data scraped from the internet, and pirated e-books[1][2][3].

[1]: https://finance.yahoo.com/news/nvidia-accused-trying-cut-dea...

[2]: https://arstechnica.com/tech-policy/2025/12/openai-desperate...

[3]: https://www.businessinsider.com/anthropic-cut-pirated-millio...


> Microcontrollers are circuits that are designed. LLMs are circuits that learned using vast amounts of data

So I suppose the AI companies employ all those data scientists and low level performance engineers to what, manage their website perhaps?

It's poor form to go around inserting your pet issue where it isn't relevant.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: