• 0 Posts
  • 36 Comments
Joined 5 months ago
cake
Cake day: June 11th, 2024

help-circle
  • I’ll grant that PHP is set up to allow some super shitty code, but on fairness to the language; WordPress is a dumpster fire (compounded by endless awful plugins). That’s compounded by it’s ubiquity, so it’s a massive target.

    I just set up mbin as a single-user instance, and other than a bug I found (that they fixed live with me, in chat, including PRs), it’s been awesome.

    I hope your instance continues to work well for you 👍




  • First, a chat bot is not an API. Second, they were talking about the the formatting and delivery method of the data, not the content.

    Regarding the output of the model: Some repos are entirely READMEs by their nature. No code, just documentation and walkthroughs. Notwithstanding that; If I set a flag that’s says “don’t use my data” and they use it anyway, that’s theft, even if it’s only one file, even if the file is just a description of the code. That’s my work, not yours. You don’t get to use it however you want, unless I specifically note that it’s public domain (or you use it and follow the license, like attributing me, or linking to the repo, etc).

    As to the difference between a bot and a human (re: stack overflow)? The former is a representative of a company (automation or not, whether it’s a bot or a page on their corporate site), the latter is a person relating experience and opinion. The legal difference is that one is using the data commercially, and the other is just a person in the world, answering another person’s question for no reason other than a desire to be helpful (and if they’re decent, attributing the source instead of claiming that they’re generating wisdom on their own).

    That last parenthetical used to be called plagiarism, by the way.






  • I’m currently avoiding silicon until more apps are compiled to work on them. My last bad experience with this was trying to run virtualbox on the host and ununtu as a guest, and it ran slow as crap because some part of virtualbox wasn’t ready for silicon yet.

    Disclaimer: I generally avoid Apple like the plague, my comment and experience are specific to a job that really wanted me to use a macbook in my role as a Linux systems admin. My specific complaint may well have been adressed literally years ago by now.


  • Agreed on all points, except my personal interpretation of “fair use” specific to the case of generative models.

    You call out “doesn’t replace the original work”. Is that not how you see an LLM Q/A bot replacing a user going to a git repo for established examples, or a website for an article (generating page views, subscriptions, ad revenue), or similar? Why would anyone go to the source materials if they’re getting their answer from the bot?

    This is practically the same as when Google started showing articles in AMP, and not bringing people to the original website, is it not?



  • The MPAA and music industry would beg to differ. As would the US courts, as well as any court in a country we share copyright agreements with.

    Consider that if a movie uses a scene from another movie without permission, or a music producer uses a melody without permission, or either of them use too much of an existing song without permission, everyone sues everyone else, and they win.

    Consider also that if a large corporation uses an individual’s content without permission, we have documented cases of the individual suing, and winning (or settling).

    Some other facts to consider;

    • An mp3 file is not inherently illegal. Nor is a torrent file/tracker/download.
    • If the mp3 file contains audio you don’t own the rights to, it is illegal, same for the torrent you used to download/distribute it. In the eyes of the law, it’s theft.
    • A trained LLM or image generation model is not inherently theft, if you only use open-source or licensed/owned content to train it
    • (at odds in our conversation) What of a model that eas trained with content the trainer didn’t own?

    In the mp3 example, its largely an individual stealing from a large company. On the Internet, this is frequently cheered as the user “sticking it to the man” (unless, of course, you’re an indie creator who can’t support yourself because everyone’s downloading your content for free). Discussions regarding the morality of this have been had - and will be had - for a long time, but it’s legality is a settled matter: It’s not legal.

    In the case of “AI” models, its large companies stealing from a huge number of individuals who have no support or established recourse.

    You’re suggesting that it’s fine because, essentially, the creators haven’t lost anything. This makes it extremely clear to me that you’ve never attempted to support yourself as a creator (and I suspect you haven’t created anything of meaning in the public domain either).

    I guess what it comes down to is this; If creators can be stolen from without consequence, what incentive does anyone have to create anything? Are you going to work your 40-60 hours a week, then come home and work another 20-40 hours to create something for no personal benefit other than the act of creation? Truely, some people will. Most wont.





  • This “fair use” argument is excellent if used specifically in the context of “education, not commercialization”. Best one I’ve seen yet, actually.

    The only problem is that perplexity.ai isn’t marketing itself as educational, or as a commentary on the work, or as parody. They tout themselves as a search engine. They also have paid “pro” and “enterprise” plans. Do you think they’re specifically contextualizing their training data based on which user is asking the question? I absolutely do not.



  • you got some criticism and now you’re saying everyone else is a bot or has an agenda

    Please look up ad hominem, and stop doing it. Yes, their responses are a distraction from the topic at hand, but so were the random posts calling OP paranoid. I’d have been on the defensive too.

    [Our company] publish[s] open source work … anyone is free to use it for any purpose, AI training included

    Great, I hope this makes the models better. But you made that decision. OP clearly didn’t. In fact, they attempted to use several methods to explicitly block it, and the model trainers did it anyway.

    I think that the anti-AI hysteria is stupid virtue signaling for luddites

    Many loudly outspoken figures against the use of stolen data for the training of generative models work in the tech industry, myself included (I’ve been in the industry for over two decades). We’re far from Luddites.

    LLMs are here

    I’ve heard this used as a justification for using them, and reasonable people can discuss the merits of the technology in various contexts. However, this is not a justification for defending the blatant theft of content to train the models.

    whether or not they train on your random project isn’t going to affect them in any meaningful way

    And yet, they did it while ignoring explicit instructions to the contrary.

    there are more than enough fully open source works to train on

    I agree, and model trainers should use that content, instead of whatever they happen to grab off every site they happen to scrape.

    Better to have your work included so that the LLM can recommend it to people or answer questions about it

    I agree if you give permission for model trainers to do so. That’s not what happened here.


  • “The world seeing [their] work” is not equal to “Some random company selling access to their regurgitated content, used without permission after explicitly attempting to block it”.

    LLMs and image generators - that weren’t trained on content that is wholly owned by the group creating the model - is theft.

    Not saying LLMs and image generators are innately thievery. It’s like the whole “illegal mp3” argument. mp3s are just files with compressed audio. If they contain copyrighted work, and obtained illegitimately, THEN their thievery. Same with content generators.


  • Eh. This is not a new argument, and not the first evidence of it. I don’t think you’re gonna be high on their list of retaliation targets, if you register at all (to say nothing of the low-to-middling reach of the fediverse in general).

    Hell, just look at photographers/painters v. image generators, or the novel/article/technical authors v. … practically all LLMs really, or any other of a dozen major stories about “AI” absorbing content and spitting out huge chunks of essentially unmodified code/writing/images.