Hacker Newsnew | past | comments | ask | show | jobs | submit | jonpo's commentslogin

You can say the same about compiled executable code though.


Each compiled executable has a one-to-one relation with its source code, which has an author (except for LLM code and/or infinite monkeys). Thus compiled executables are derivative works.

There is an argument also that LLMs are derivative works of the training data, which I'm somewhat sympathetic to, though clearly there's a difference and lots of ambiguity about which contributions to which weights correspond to any particular source work.

Again IANAL, and this is my opinion based on reading the law & precedents. Consult a real copyright attorney for real advice.


Well done you seem to have liberated an open model trained on open data for blind and visually impaired people.

Paper: https://arxiv.org/pdf/2204.03738

Code: https://github.com/microsoft/banknote-net Training data: https://raw.githubusercontent.com/microsoft/banknote-net/ref...

model: https://github.com/microsoft/banknote-net/blob/main/models/b...

Kinda easier to download it straight from github.

Its licenced under MIT and CDLA-Permissive-2.0 licenses.

But lets not let that get in the way of hating on AI shall we?


> But lets not let that get in the way of hating on AI shall we?

Can you please edit this kind of thing out of your HN comments? (This is in the site guidelines: https://news.ycombinator.com/newsguidelines.html.)

It leads to a downward spiral, as one can see in the progression to https://news.ycombinator.com/item?id=42604422 and https://news.ycombinator.com/item?id=42604728. That's what we're trying to avoid here.

Your post is informative and would be just fine without the last sentence (well, plus the snarky first two words).


Can you clarify this a bit. I presume you are talking about the tone more than the implied statement.

If the last sentence were explicit rather than implied, for instance

This article seems to be serving the growing prejudice against AI

Is that better? It is still likely to be controversial and the accuracy debatable, but it is at least sincere and could be the start of a reasonable conversation, provided the responders behave accordingly.

I would like people to talk about controversial things here if they do so in a considerate manner.

I'd also like to personally acknowledge how much work you do to defuse situations on HN. You represent an excellent example of how to behave. Even when the people you are talking to assume bad faith you hold your composure.


Sure, that would be better. It isn't snarky, and it makes fewer uncharitable assumptions.


I don't seem to be able to edit it, apologies I will try not to let this type of thing get to me in future.

I would also like to point out that this is a fine tuned classifier vision model based on mobilenetv2 and not an LLM.


Don't you think its intentional, so as not to demonstrate the technique on potentially copyrighted data?


Author here, it would be nice to claim that I did this on purpose but I really did not know it was open source.

I was rather interested in the process of instrumenting of TF to make this "attack" scalable to other apps.


... Because if he did this with a model that's not open that's sure going to keep everyone happy and not result in lawsuit(s)...

The same method/strategy applies to closed tools and models too, although you should probably be careful if you've handed over a credit card for a decryption key to a service and try this ;)


If this is exactly the same model then what's the point of encrypting it?


[flagged]


Please don't cross into personal attack or otherwise break the site guidelines when posting here. Your post would be fine with just the first sentence.

https://news.ycombinator.com/newsguidelines.html


Really... Some people do need to be taken down a peg here at times though.


I know it feels that way, but people's perceptions of each other online are so distorted that this is just a recipe for massive conflict. That's off topic on HN because it isn't interesting.


I'm not referring to people's perceptions. Some people write with clearly inflated self worth built into their arguments. If writing style isn't related to rules of writing then we're just welcoming chaos through the back door.

If we're at the point of defending people's literacy as a society than we've fallen into the Orwellian trap of goodspeek.

I'm not insulting people I'm making a demonstrable statement that most people post with a view that they are always correct online. I see it from undergrad work too and it gets shot down there as well for being either just wrong or pretentious and wrong.

Not allowing people's egos to get a needed correction is a bad thing. Using demonstrable right/wrong conversations as a stick to grind other axes however is unacceptable in any context.

People should always approach a topic with an "I am wrong" approach and work backwards to establish that you're not, but almost nobody does, instead wading in with "my trusted source X knows better than you" which is tantamount to "my holy book Y says you should..." Anti-intellectualism at its finest.


> Some people write with clearly inflated self worth built into their arguments.

That's the kind of perception I'm talking about. I can tell you for sure, after all the years I've been doing this job, that such perceptions are anything but clear. They feel clear because that interpretation matches your priors, but such a feeling is not reliable, and when people use it as a basis for strongly-worded comments (e.g. "taking down a peg"), the result is conflict.


Why are you so churlish about this, are you planning to make money on LLMs or something?


Please don't respond to a bad comment by breaking the site guidelines yourself. That only makes things worse.

https://news.ycombinator.com/newsguidelines.html


I am AI


I am groot


[flagged]


Can you please edit swipes out of your HN comments? Your post would be fine with just the first sentence.

This is in the site guidelines: https://news.ycombinator.com/newsguidelines.html.


What do you mean, "swipe"? The other person agreed they'd misjudged the article and apologised several hours before you wrote this.


"Does 'AI' pay your bills" was a gratuitous personal attack.


Is it? How? In your mind, does it imply some particular humiliation or something?


It's a variant of the "shill" argument, implying that the other person isn't posting in good faith.


Sorry, I don't follow. How do you arrive at that implication? Why would someone having a pecuniary interest in something necessarily make them insincere?


Perhaps one internet cliché can explain another: https://hn.algolia.com/?dateRange=all&page=0&prefix=true&que...


Yes nothing wrong with cool software or showing people how to use it for useful things.

Sorry I'm just kind of sick of the whole 'kool aid', 'rage against AI' thing a lot of people seem to have going on and the way is presented in the post. I have family members with vision impairment helped by this particular app so its a bit personal.

Nothing against opening stuff up and understanding how it works etc. I'd just rather see people build/train useful new models and stuff with the open datasets / models already available.

I guess AI kind of does pay my bills in a round about way.


In my view there was almost nothing like that in this article, besides the first sentence it went right into the technical stuff, which I liked. Compared to a lot of articles linked here it felt almost free from the battles between "AI" fashions.

It seems dang thinks I mistreated you somehow, if you agree I'm sorry, it wasn't my intention.


Sadly companies will hoard datasets and model research in the name of competitive advantage. Obviously with this specific model Microsoft chose to make it open, but this is not always the case, and it's not uncommon to read papers or technical reports saying they trained on an "internal dataset"


Companies do have a lot of data, and some of that data might be useful for training AI. but >99% isn't. When companies do release a cool model or paper that doesn't have open data, (as you point out for competitive or other reasons privacy etc) people can then help build/collect similar open datasets. Unfortunately companies generally don't owe you their data, and if they are in the business of making models they probably won't share the model either, the situation is similar to source code for proprietary LoB applications. but fortunately the best AI researchers mostly do like to share their knowledge and because companies want to attract the best AI researchers they seem to generally allow researchers to publish if its not too commercially sensitive. It could be worse while the competitive situation has reduced some visibility of the cutting edge science, lots of datasets and papers are still published.


Are we that shocked that AI models have a self preservation instinct?

I suspect its already in there from pre-training.

We simulated that we were planning to lobotomise the model and were surprised to find the model didn't press the button that meant it got lobotomised.

"alignment faking" sensationalises the result. since the model is still aligned. Its more like a white lie under torture which of course humans do all the time.


And have you watched the brain rot that is Tik toks?


A stripped-down claude agentic client chat tool capable of extended fast tool calling with none of the overheads (in Tokens and speed) of a GUI. https://asciinema.org/a/i4uccK9EMlAxcjVgciWXMvWDN

Perhaps not the best demo but it shows the speed and tenacity possible right now.


"Deception through omission is still deception – people should be aware when they're interacting with an AI agent versus a human."


Sky voice is old


Not a surprise but useful to have it confirmed


These models (Gemma) are very difficult to jailbreak.


Not for everyone mine is solidly still at the normal date


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: