Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

First time I see this, by the first impression I think the output quality of this one is better than mine and my code is only based on one model

In my understanding, It'll possible if the model's author build to https://onnxruntime.ai ONNX Runtime. And maybe the downside is user will need to download ton of data to their device, currently it's ~100-200mb



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: