Hacker Newsnew | past | comments | ask | show | jobs | submit | freeatnet's commentslogin

Interesting! A friend recently asked me if I knew of any tools to improve GPU observability across their deployments (primarily for cost tracking purposes, I think), but he was looking for an OSS solution. Do you plan to open source this?


We have considered this and may go down this route in the future. One thing we asked ourselves was what open sourcing provides. Usually it's a desire for privacy or cost in the form of self-hosting, among other reasons.

Currently, our free version is self-hosted and monitors clusters with up to 64 GPUs. We feel this will work for many use cases, especially just to try it out. Monitoring GPUs typically requires you to deploy something where your GPUs live. Since you’re already installing software on your cluster, you might as well keep your data there too.


Your Github repo says you need 120 GB of persistent storage, but our bare metal GPU clusters only have local storage. Would like to try your thing, but hosting the data with the GPUs is a pretty big blocker for us.


Ahh yes...here’s how you solve that. Just install the Neurox Control plane onto any regular Kubes cluster (doesn’t need GPUs, just needs persistent storage. ie: EKS, AKS, GKE, etc) without that last flag in the instructions: `--set workload.enabled=true` (<-- leave this out). More info: https://docs.neurox.com/installation/alternative-install-met...

Then on your GPU cluster w/o disk, you just need to install the Neurox Workload agent. In the Web Portal UI, click on Clusters > New Cluster and copy/paste the snippet there.


Oh sweet, I'll take a look. Thanks!


Does anyone know of a good comparison of these bundling methods?


That's a pretty cool approach! Alas, the interface seems to have gotten stuck on step 9 for me.


Thanks, yeah. I wanted to see what HN thought before I took the time to finish it. Looks like I will.


One more vote of support. I'm reading it now.

I've always wished that Kindle had this. One can buy a Kindle translation (and some popular books have communities around translations: https://tinwhiskers.net/post/harry_potter_spanish, https://forum.duolingo.com/comment/9229774/Harry-Potter-in-S...) as well as a translating dictionary, but one must find a book that's very close to the reader's level or reading will be impossibly slow (highlight word, click to translate, repeat). Your contextual translations and next/play button are way faster.


Good start. I think it would improve with some more layout & colour scheme work. Keep at it!


thanks !


How do you plan to make money?


Also: how are you going to make this app _the_ way people do group scheduling?


What differentiates you from Reddit or any other community-sourced link sites?


It'd be cool if it didn't suggest names that very obviously don't sound good in (at least common) non-english languages, eg., I got "culojoo" as a suggestion for 3 syllables, which won't sound good in Spanish at all.


Names are totally random, not checked against any dictionary, that's what I thought was fun. It would be hard to appreciate the goodness of a name, at least programmatically, how would you do that ?


tenken is right, I meant to say "not let names have vulgar connotations in other languages." In the example above, the name started with "culo", which translates into "ass" in Spanish. Here's an example of a service that does that: http://wordsafety.com/

Nevertheless, it's a pretty cool idea! :)


He's saying be sure the generator doesn't sound like a bad name in any language.

I'd say make a dictionary of bad words and their syllble sounds. Then disallow close Levienstein based words....


Levenstein is an absolute metric, I think something like soreson-dice would be more useful.

Regardless, if you take the short keywords and blacklist them by approximation with curse words from several languages I think it would be really hard to get something at all.


I feel this could benefit from some additional context.


hey! see reply above^^ :)


Appears to be down at the moment. Would have been neat to be able to check it out.


Sorry about that. It is running on heroku free instance (limited to 18 hours per day). It should be back online in less than an hour.


It should be working now


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: