• 1 Post
  • 71 Comments
Joined 2 years ago
cake
Cake day: July 9th, 2023

help-circle
  • the most generous reading of your arguments is that you are philosophically defeatist

    That’s probably a fair assessment.

    But I feel like the core of my argument remains: I’m not disputing that MS or Google or Amazon or Apple services are sold to people and orgs who use them to commit evil. Of course they are.

    But these aren’t munitions. They are general-purpose computing products being turned to evil outcomes by bad actors. The article, for example, cites Microsoft’s open-source LAVENDER, which is a general purpose image and video analysis tool for AI. Describing it as:

    ‘Lavender’, an AI-powered system designed to identify bombing targets

    This simply isn’t true. Somebody in the Israeli military used LAVENDER to process video data to identify bombing targets, like somebody might use a hammer to smash someone’s head in. The articles you cite are full of rhetorical tricks to imply that Microsoft corporate had some hand in the decision making, but it’s genuinely all “well the Israeli military has some Azure servers, therefore Microsoft killed people”.

    Which militaries should Microsoft (or Google or Apple or Amazon, etc) be allowed to sell products to? Who makes that determination? A cohort of employees or consumers? NGOs?

    If government makes the call – distilling a public consensus on the matter, one hopes – then I can see some reasonable way to approach this question.

    EDIT: Details on LAVENDER:

    https://www.microsoft.com/en-us/research/publication/lavender-unifying-video-language-understanding-as-masked-language-modeling/



  • Honestly, I struggle to draw a connection between world conflict and non-military technology like Windows or cell phones or whatever.

    Is every single Israeli resident complicit in what their government is doing? None of them should be allowed to use Windows? What about Israelis outside of Israel? What about people who support Israel? What about (gasp) Jews? How do you even enforce any of this without massive overreach by the companies?

    Call on Microsoft or Apple all you want, ultimately I don’t think a company should ban sales to customers on the argument that those customers might not have morals aligned to the company. Not that it’s even possible, with world supply chains being what they are.



  • With respect to the article, it’s wrong. AI help desk is already a thing. Yes, it’s terrible, but human help desk was already terrible. Businesses are ABSOLUTELY cutting out tier 1 call center positions.

    LLMs are exceptionally good at language translation, which should be no surprise as that kind of statistical chaining is right up their alley. Translators are losing jobs. AI Contract analysis & legal blacklining are going to put a lot of junior employees and paralegals out of business.

    I am very much an AI skeptic, but I also recognize that people who do the things LLMs are already pretty good at are in real trouble. As AI tools get better at more stuff, that target list of jobs will grow.













  • it’s basically impossible to tell where parts of the model came from

    AIs are deterministic.

    1. Train the AI on data without the copyrighted work.

    2. Train the same AI on data with the copyrighted work.

    3. Ask the two instances the same question.

    4. The difference is the contribution of the copyrighted work.

    There may be larger questions of precisely how an AI produces one answer when trained with a copyrighted work, and another answer when not trained with the copyrighted work. But we know why the answers are different, and we can show precisely what contribution the copyrighted work makes to the response to any prompt, just by running the AI twice.



  • There is literally not one single piece of art that is not derived from prior art in the past thousand years.

    This is false. Somebody who looks at a landscape, for example, and renders that scene in visual media is not deriving anything important from prior art. Taking a video of a cat is an original creation. This kind of creation happens every day.

    Their output may seem similar to prior art, perhaps their methods were developed previously. But the inputs are original and clean. They’re not using some existing art as the sole inputs.

    AI only uses existing art as sole inputs. This is a crucial distinction. I would have no problem at all with AI that worked exclusively from verified public domain/copyright not enforced and original inputs, although I don’t know if I’d consider the outputs themselves to be copyrightable (as that is a right attached to a human author).

    Straight up copying someone else’s work directly

    And that’s what the training set is. Verbatim copies, often including copyrighted works.

    That’s ultimately the question that we’re faced with. If there is no useful output without the copyrighted inputs, how can the output be non-infringing? Copyright defines transformative work as the product of human creativity, so we have to make some decisions about AI.