GioCities

blogs by Gio

Tagged: ai

⚖ The ambiguous "use"

I keep seeing people make this error, especially in social media discourse. Somebody wants to “use” something. Except obviously, it’s not theirs, and so it’s absurd for them to make that demand, right?

Quick examples§

I’m not trying to pick on this person at all: they’re not a twitter main character, they’re not expressing an unusual opinion here, they seem completely nice and cool. But I think this cartoon they drew does a good job of capturing this sort of argument-interaction, which I’ve seen a lot:

I’ve also seen the exact inverse of this: people getting upset at artists because once the work is “out there” anyone should be able to “use” it. (But I don’t have a cartoon of this.)

There is an extremely specific error being made in both cases here, and if you can learn to spot it, you can save yourself some grief. What misuse is being objected to? What are the rights to “certain things” being claimed?

The problem is that “use” is an extremely ambiguous word that can mean anything from “study” to “pirate” to “copy and resell”. It can also cover particularly sensitive cases, like creating pornography or editing it to make a political argument.

webcomicname: beliefs you do not agree with

But everything people do is “using” something. By itself, “use” is not a meaningful category or designation. Say you buy a song — listening to it, sampling it, sharing it, performing it, discussing it, and using it in a video are all “uses”, but the conversations about whether each is appropriate or not are extremely distinct. If you have an objection, it matters a lot what specific use you’re talking about.

But if you’re not specific, there are unlimited combinations of “uses” you could be talking about, and you could mean any of them. And when people respond, they could be responding to any of those interpretations. There’s no coherent argument in any sweeping statement about “use”; the only things being communicated are frustration and a team-sports-style siding with either “artists” or “consumers” (which is a terrible distinction to make!).

Formal logic§

This is not a new problem. This is the Fallacy of Equivocation, which is a subcategory of Fallacies of Ambiguity. This is when a word (in this case, “use”) has more than one meaning, and an argument uses the word in such a way that the entire position and its validity hinge on which definition the reader assumes.

The example of this that always comes to my mind first is “respect”, because this one tumblr post from 2015 said it so well:

flyingpurplepizzaeater Sometimes people use “respect” to mean “treating someone like a person” and sometimes they use “respect” to mean “treating someone like an authority”

and sometimes people who are used to being treated like an authority say “if you won’t respect me I won’t respect you” and they mean “if you won’t treat me like an authority I won’t treat you like a person”

and they think they’re being fair but they aren’t, and it’s not okay.

See, here the “argument” relies on implying a false symmetry between two clauses that use the same word but with totally different meanings. And, in disambiguating the word, the problem becomes obvious.

Short-form social media really exacerbates the equivocation problem by encouraging people to be concise, which leads to accidental ambiguity. But social media also encourages people to take offense at someone else being wrong as the beginning of a “conversation”, which encourages people to use whatever definition of other people’s words makes them the wrongest.

Since I’m already aware that copyright is a special interest of mine, I try to avoid falling into the trap of modeling everything in terms of copyright by default, Boss Baby style. But this is literally the case of a debate over who has the “right” to various “uses” of things that are usually intangible ideas, so I think it’s unavoidably copyright time again.

95% of the time when people complain about a “use” being inappropriate, or complain about people feeling “entitled” to a “use”, they’re talking about a specific kind of use that has its own field of study under the broad category of “intellectual property”.

When people categorically object to “use” of their work, this can mean one of two things. One is that they subscribe to the extremely juvenile idea that authors have an unlimited, perpetual, inalienable right to control all publication, derivation, and interpretation of work “they created”, for some definition of “create”. This is wrong. The question of “what rights do people have over information they produce” is an incredibly complex topic, and the answer isn’t “all of them, stupid”.

The other case — and this is usually the case — is that they mean something much more specific by the word “use”, and have simply failed to explain it. Maybe they specifically mean redistribution, or plagiarism, or a more nebulous idea of “ripping off”, or some combination of those things. Maybe you agree with this, or maybe you don’t, you don’t even know. You can’t possibly talk about it until they define what their assertion is!

This means in arguments about rights to “use” work which refuse to specify the exact rights they mean, the “sides” are effectively talking past each other. In good faith, this can just be a mistake, but in bad faith can be deliberate outrage farming. Unfortunately, this turns out to be a very effective sleight of hand when used deliberately.

It is very easy for someone to “feel” wronged when someone else is “using” “their” “thing”. But, if they haven’t actually been wronged, they’ll find it dificult to articulate why there’s a legitimate grievance, because they don’t have one. When this happens, instead of reconsidering their own emotions, people often cover up their own error by pretending they don’t need to articulate their grievance at all, and keep all their complaints convincingly ambiguous.

This happened with the Internet Archive lawsuit, an ugly saga I’ve discussed in detail. The rhetorical argument against the archive was this wiggly little “use”: “they’re using our books without paying us!” When in reality, the way the Internet Archive’s library was “using” them was lending out books they owned legally. But “we’re demanding libraries stop lending our books” makes you sound obviously evil, whereas “you have to pay us when you use our IP” doesn’t.

A person who believes “artists should be paid when you use their work” might also agree that “the builder doesn’t get paid every time the house sells” and “artists should study books on technique”. But the second two statements both conflict with the first one! How can they all make sense at once? Because the first statement uses a wiggly “use” that folds itself down to be as reasonable as you want it to be.

AI “use”§

I have another piece I’m writing about the very complicated relationship LLM AI has with “using” work. I’ll talk about this much more in-depth then, but here are some early thoughts to chew on. People really like using the ambiguous “use” with regards to AI, and I think one of the reasons is to create an outrage that’s amplified by the ambiguous metaphor.

Take the article1 OpenAI Pleads That It Can’t Make Money Without Using Copyrighted Materials For Free. Only in the body of the piece does it explain that the actual conflict is the unsettled question of whether or not AI training — an act that is mechanically very different from copying — is considered to be a violation of copyright that needs specific licensing.

This is a pretty detailed and technical question, actually, and one I have an upcoming major essay dedicated to. It’s also an important question! It deserves to be considered carefully and given a thoughtful, specific answer. Simplifying it down to “use” isn’t just inaccurate, it’s intellectually lazy.

An accurate summary would be “OpenAI argues Training on Copyrighted Materials Is Non-Infringing Use”, which is the thing that happened. But “OpenAI can’t make money without using copyrighted materials for free” bakes the outrage directly into the headline, and creates an easily-digestible narrative for people: OpenAI is profiting by exploiting a resource it should be paying for. But that all hinges on the ambiguous “use” in the title.

Same thing with News Corp sues Perplexity for ripping off WSJ and New York Post by Emma Roth2, where the imprecise phrase “ripping off” lets the reader immediately stop thinking about the problem. What does “ripping off” mean? It sure sounds bad when you don’t explain what happened. Was it plagiarism? Direct copy-paste? Why does News Corp think it was wronged?

In reality, Perplexity made a website that told people what another website said. Some people will agree with News Corp that telling users what another website says is an offensive misappropriation, but other people will argue that describing the content of other sites should be allowed speech. Both positions are interesting! But by using the vague phrase “ripping off” instead of describing the situation to people, this headline is actively obfuscating the question at hand and instead focusing on the “feeling” of offense. Framing the story as “News Corp feels offended” instead of describing the event in question turns the conversation into the team sports “who do you like more” contest, which is worthless.

Conclusion§

Don’t talk past each other!

Encourage an understanding of the topics you care about!

Don’t reduce complex questions down to team sports!

Precision of language!


  1. this is another xerox-of-a-xerox article; it’s mostly just an unnecessary summary of actual reporting by The Telegraph. 

  2. Hey, look at that byline, she’s another crypto skeptic who naturally pivoted to AI skepticism. We just talked about that! 

đŸ–± Is AI eating all the energy? Part 2/2

  • Posted in cyber

Part 2: Growth, Waste, and Externalities§

The AI tools are efficient according to the numbers, but unfortunately that doesn’t mean there isn’t a power problem. If we look at the overall effects in terms of power usage (as most people do), there are some major problems. But if we’ve ruled out operational inefficiency as the reason, what’s left?

The energy problems aren’t coming from inefficient technology, they’re coming from inefficient economics. For the most part, the energy issues are caused by the AI “arms race” and how irresponsibly corporations are pushing their AI products on the market. Even with operational efficiency ruled out as a cause, AI is causing two killer energy problems: waste and externalities.

đŸ–± Is AI eating all the energy? Part 1/2

  • Posted in cyber

Recent tech trends have followed a pattern of being huge society-disrupting systems that people don’t actually want. Worse, it then turns out there’s some reason they’re not just useless, they’re actively harmful. While planned obsolescence means this applies to consumer products in general, the recent major tech fad hypes — cryptocurrency, “the metaverse”, artificial intelligence
 — all seem to be comically expensive boondoggles that only really benefit the salesmen.

Monorail!

The most recent tech-fad-and-why-it’s-bad pairing seems to be AI and its energy use. This product-problem combo has hit the mainstream as an evocative illustration of waste, with headlines like Google AI Uses Enough Electricity In 1 Second To Charge 7 Electric Cars and ChatGPT requires 15 times more energy than a traditional web search.

It’s a narrative that’s very much in line with what a disillusioned tech consumer expects. There is a justified resentment boiling for big tech companies right now, and AI seems to slot in as another step in the wrong direction. The latest tech push isn’t just capital trying to control the world with a product people don’t want, it’s burning through the planet to do it.

But, when it comes to AI, is that actually the case?

What are the actual ramifications of the explosive growth of AI when it comes to power consumption? How much more expensive is it to run an AI model than to use the next-best method? Do we have the resources to switch to using AI on things we weren’t before, and is it responsible to use them for that? Is it worth it?

These are really worthwhile questions, and I don’t think the answers are as easy as “it’s enough like the last thing that we might as well hate it too.” There are proportional costs we have to weigh in order to make a well-grounded judgement, and after looking at them, I think the energy numbers are surprisingly good, compared to the discourse.

đŸ–± Reddit: Your API *IS* Your Product

  • Posted in cyber

Reddit is going the same route as Twitter by making “API access” prohibitively expensive. This is something they very famously, very vocally said they would not do, but they’re doing it anyway. This is very bad for Reddit, but what’s worse is it’s becoming clear that companies think that this is a remotely reasonable thing to do, when it’s very critically not.

It’s the same problem we see with Twitter and other late-capitalist hell websites: Reddit’s product is the service it provides, which is its API. The ability for users to interact with the service isn’t an auxiliary premium extra, it’s the whole caboodle!

I’ll talk about first principles first, and then get into what’s been going on with Reddit and Apollo. The Apollo drama is very useful in that it directly converts the corporate bullshit that sounds technical enough to make sense into something very easy to understand: a corporation hurting them, today, for money.

The API is the product§

Reddit and all these other companies who are making user-level API access prohibitively expensive have forgotten that the API is the product. - The API is the interface that lets you perform operations on the site. The operations a user can do are the product, they’re not auxiliary to it!

“Application programming interface” is a very formal, internal-sounding term for a system that is none of those things. The word “programming” in the middle comes from an age where using a personal computer at all was considered “programming” it.

What an API really is a high-level interface to the web application that is Reddit. Every action a user can take — viewing posts, posting, voting, commenting — goes from the app (which interfaces with the user) to the API (which interfaces with the Reddit server), gets processed by the server using whatever-they-use-it-doesn’t-matter, and the response is sent back to the user.

The API isn’t a god mode and it doesn’t provide any super-powers. It doesn’t let you do anything you can’t do as a user, as clearly evidenced by the fact that all the actions you do on the Reddit website go through the API too.

The Reddit website, the official Reddit app, and the Apollo app all interface with the user in different ways and on different platforms, but go through the same API to interact with what we understand as “Reddit”. The fact that the API is the machine interface without the human interface should also concisely explain why “API access” is all Apollo needs to build its own app.

Right now, you can view the announcement thread at https://www.reddit.com/r/apolloapp/comments/144f6xm/apollo_will_close_down_on_june_30th_reddits/, and you can view the “API” data for the same thread at https://www.reddit.com/r/apolloapp/comments/144f6xm/apollo_will_close_down_on_june_30th_reddits.json. It’s not very fun to look at, but it’s easy to tell what you’re looking at: the fundamental representation of the page without all the trappings of the interface.

Public APIs are good for both the user and the company. They’re a vastly more efficient way for people to interact with the service than by automating interaction (or “scraping”). Having an API cuts out an entire layer of expense that, without an API, Reddit would pay for.

The Reddit service is the application, and you interface with it through WHATEVER. Whatever browser you want, whatever browser extensions you want, whatever model phone you want, whatever app you want. This is fundamentally necessary for operability and accessibility.

The API is the service. The mechanical ability to post and view and organize is what makes Reddit valuable, not its frontend. Their app actually takes the core service offering and makes it less attractive to users, which is why they were willing to pay money for an alternative!

đŸ–± So you want to write an AI art license

  • Posted in cyber

Hi, The EFF, Creative Commons, Wikimedia, World Leaders, and whoever else,

Do you want to write a license for machine vision models and AI-generated images, but you’re tired of listening to lawyers, legal scholars, intellectual property experts, media rightsholders, or even just people who use any of the tools in question even occasionally?

You need a real expert: me, a guy whose entire set of relevant qualifications is that he owns a domain name. Don’t worry, here’s how you do it:

Given our current system of how AI models are trained and how people can use them to generate new art, which is this:

CurioModelAliceCurioModelAliceHello. Here are N images andtext descriptions of what they contain.Training (looks at images, "makes notes", discards originals)OK. I can try to make similar images from my notes,if you tell me what you want.Hello. I would like a depiction of this new thing you've never seen before.OK. Here are some possibilites.

đŸ–± Replika: Your Money or Your Wife

  • Posted in cyber

If1 you’ve been subjected to advertisements on the internet sometime in the past year, you might have seen advertisements for the app Replika. It’s a chatbot app, but personalized, and designed to be a friend that you form a relationship with.

That’s not why you’d remember the advertisements though. You’d remember the advertisements because they were like this:

Replika "Create your own AI friend" "I've been missing you" hero ad

Replika ERP ad, Facebook (puzzle piece meme) Replika ERP ad, Instagram

And, despite these being mobile app ads (and, frankly, really poorly-constructed ones at that) the ERP function was a runaway success. According to founder Eugenia Kuyda the majority of Replika subscribers had a romantic relationship with their “rep”, and accounts point to those relationships getting as explicit as their participants wanted to go:

erp1

So it’s probably not a stretch of the imagination to think this whole product was a ticking time bomb. And — on Valentine’s day, no less — that bomb went off. Not in the form of a rape or a suicide or a manifesto pointing to Replika, but in a form much more dangerous: a quiet change in corporate policy.

Features started quietly breaking as early as January, and the whispers sounded bad for ERP, but the final nail in the coffin was the official statement from founder Eugenia Kuyda:

“update” - Kuyda, Feb 12 These filters are here to stay and are necessary to ensure that Replika remains a safe and secure platform for everyone.

I started Replika with a mission to create a friend for everyone, a 24/7 companion that is non-judgmental and helps people feel better. I believe that this can only be achieved by prioritizing safety and creating a secure user experience, and it’s impossible to do so while also allowing access to unfiltered models.

People just had their girlfriends killed off by policy. Things got real bad. The Replika community exploded in rage and disappointment, and for weeks the pinned post on the Replika subreddit was a collection of mental health resources including a suicide hotline.

Resources if you're struggling post

Cringe!§

First, let me deal with the elephant in the room: no longer being able to sext a chatbot sounds like an incredibly trivial thing to be upset about, and might even be a step in the right direction. But these factors are actually what make this story so dangerous.

These unserious, “trivial” scenarios are where new dangers edge in first. Destructive policy is never just implemented in serious situations that disadvantage relatable people first, it’s always normalized by starting with edge cases and people who can be framed as Other, or somehow deviant.

It’s easy to mock the customers who were hurt here. What kind of loser develops an emotional dependency on an erotic chatbot? First, having read accounts, it turns out the answer to that question is everyone. But this is a product that’s targeted at and specifically addresses the needs of people who are lonely and thus specifically emotionally vulnerable, which should make it worse to inflict suffering on them and endanger their mental health, not somehow funny. Nothing I have to content-warning the way I did this post is funny.

Virtual pets§

So how do we actually categorize what a replika is, given what a novel thing it is? What is a personalized companion AI? I argue they’re pets.