GioCities

blogs by Gio

Tagged: ip

politics Netflix's Big Double-Dip

Netflix is finally turning the screws on multi-user accounts. That “finally” is exasperation in my voice, not relief. Netflix is demanding you pay them an extra surcharge to share your account with remote people, and even then caps you at paying for a maximum of two. It’s been threatening to do something like this for a long, long time:

Since 2011, when the recording industry started pushing through legal frameworks to criminalize multi-user account use by miscategorizing “entertainment subscription services” as equivalent to public services like mail, water, and electricity for the purposes of criminal prosecution,

Since similar nonsense in 2016 exploiting the monumentally terrible Computer Fraud and Abuse Act,

Since 2019, when Netflix announced (to its shareholders) that it was looking for ways to limit password sharing,

Since 2021, when Netflix started tracking individual users by location and device within a paying account,

Since 2022, when it started banning group use in Portugal, Spain, and New Zealand, to disastrous consequence. Also, Canada, but temporarily. And, of course, then threatened to “crack down” on “password sharing” in “Early 2023”,

Since January, when it threatened to roll out “paid password sharing” in the “coming months”,

Since February, when it released a disastrous policy banning password sharing, then lied about the policy being an error and made a big show of retracting it due to the massive backlash, but then went ahead and did it in Canada anyway,

And finally now since just now, as it’s finally, really, for-realsies banning password sharing this quarter.

Netflix threatening this for so long was a mistake on its part, because that’s given me a long, long time for these thoughts to slowly brew in the back of my head. And there’s a lot wrong here.

the teat one this is a real graphic Netflix made!

Netflix’s pricing model🔗

So, first, what are multi-user accounts in the first place, and how does “password sharing” relate to that?

cyber So you want to write an AI art license

  • Posted in cyber

Hi, The EFF, Creative Commons, Wikimedia, World Leaders, and whoever else,

Do you want to write a license for machine vision models and AI-generated images, but you’re tired of listening to lawyers, legal scholars, intellectual property experts, media rightsholders, or even just people who use any of the tools in question even occasionally?

You need a real expert: me, a guy whose entire set of relevant qualifications is that he owns a domain name. Don’t worry, here’s how you do it:

This is an extremely condensed set of notes, designed as a high-level overview for thinking about the problem

Given our current system of how AI models are trained and how people can use them to generate new art, which is this:

sequenceDiagram
    Alice->>Model: Hello. Here are N images and<br>text descriptions of what they contain.
    Model->>Model: Training (looks at images, "makes notes", discards originals)
    Model->>Alice: OK. I can try to make similar images from my notes,<br>if you tell me what you want.
    Curio->>Model: Hello. I would like a depiction of this new <br>thing you've never seen before.
    Model->>Curio: OK. Here are some possibilites.

The works🔗

The model and the works produced with the model are both distinct products. The model is more like processing software or tooling, while the artistic works created with the model are distinctly artistic/creative output.

Models do not keep the original images they were trained on in any capacity. The only keep mathematical notes about their properties. You (almost always) cannot retrieve the original image data used from the model after training.

sequenceDiagram
    Curio->>Model: Send me a copy of one of the images you were trained on
    Model->>Curio: Sorry, I do not remember any of them exactly,<br>only general ideas on how to make art.

There is a lot of misinformation about this, but it is simply, literally the case that a model does not include the training material, and cannot reproduce its training material. While not trivial (you can’t have a model if you can’t train it at all), when done properly, the specific training data is effectively incidental.

AI-generated art should be considered new craftsmanship — specifically, under copyright law, it is new creative output with its own protections — and not just a trivial product of its inputs.

Plagiarism🔗

The fact that AI art is new creative output doesn’t mean AI art can’t be plagiarism.

Just like with traditional art, it’s completely possible for specific products to be produced to be copies, but that doesn’t make that the case for all works in the medium. You can trace someone else’s artwork, but that doesn’t make all sketches automatically meritless works.

The inner workings of tools used in the creation of an artistic work are not what determines if a given product is plagiarism, or if it infringes on a copyright. Understanding the workings of the tool can be used in determining if a work is an infringement, but it is not the deciding factor.

To use a trivial example, if I copy an image to use in an advertisement, the copyright violation is in the use of the material, and the fact that the material is, in practice, a replica of existing copyrighted work. The “copy” program isn’t the infringement, it just informs our understanding of the infringement. Monkeys on typewriters can make something that infringes copyright too.

Is using an AI model as a step in the artistic process prima facie sufficient evidence that any work generated by it is an infringement of someone else’s copyright? The answer — based on an understanding of the tools and the range of the output space — is no.

Like all new and more efficient tools, AI art tools can be used to efficiently create new work more efficiently or copy old work efficiently. Both of those cases worry certain groups, but the fact is the technology can both create new work and copy existing work.

Don’t break everything🔗

It would be monumentally terrible for the general “right for someone to use their experience of a published work” to be codified as an idiosyncratic property right that is assumed to be reserved to the copyright holder unless they specifically license it out.

Using “an experience of a published work of art to infer what art looks like” is exactly how the AI model training that people are worried about works, and that model training runs as a user-agent, so an attempt to differentiate “tool-assisted learning” from “unassisted human learning” is also a dangerous avenue. (I reject the idea that there is a meaningful distinction between “natural” and “technologically assisted” human action, in favour of network theory.)

Creating implicit or explicit “style rights” that would give artists/companies/rightsholders legal leverage against people (AI assisted or otherwise) who make works that “feel similar”, even if aspects like the subject are materially different from anything the rightsholder has copyright to, is an even-worse-but-still monumentally terrible idea.

Possibly good goals🔗

So what do actual AI artists (like the fine folks over at the AWAY collective) want to see in copyright? I think the following are safe to describe as goals:

  • Ensuring that artists — both “traditional” and tool-assisted — are free to create and share their work without endangering themselves in the process.
  • Preventing the mass-replacement of traditional artists with systems that output cheap, mass-produced works, especially if those works are derived in part from the artists this system harms.
  • Preventing a fear-induced expansion of copyright that creates new rights that ultimately only benefit corporations that stockpile the new rights and use them against artists, the way music sampling rights work today.

These seem at odds with each other.🔗

How can you retain meaningful control over your work if making it publicly visible on the internet grants corporations rights over most of its value? How can copyright distinguish between what we consider “constructive” educational use of public information (human education, as the most trivial example) and uses we would see as exploitative, like training an AI on the works made by a particular author in order to produce facsimiles of their work without compensating the original artist?

I believe mass and corporate use of AI-generated work exploiting the creative output of humans is a real danger in a way that individual artists using AI for individual works isn’t. But how do we make that distinction in a meaningful way within the framework of copyright? What, specifically, is the distinction that makes the former a serious threat to the wellbeing of both real humans and the creative market, but the latter actively beneficial to the artistic community?

The distinction cannot simply be “commercial” use, because restrictions on commercial use penalize the independent artist as much as the would-be exploiter. An artist (again, tool-assisted or human) needs to retain creative rights over their work and be able to sell it without being permanently indentured to their educators.

Nor should it be based on some arbitrary threshold like the income of the artist, or their incorporation status. Those are empty distinctions; that’s fitting the available data points to the “model” of how I feel the world should look instead of drilling down and finding what the real distinction is.

This is a hard problem, and not one I’ve solved (yet). The above are some thoughts I’ve been chewing on — I have another article I’m working on where I go into more detail on that. But there are some moves that seem like clear steps in the right direction, like licenses incorporating Creative Commons-style share-alike principles.

Possibly good ideas🔗

Licensing models (“understandings” of art) with a requirement that art generated using that model must be attributed back to the model (and, transitively, the model’s source information) is probably a good idea and something that people (model-creators) should be able to do if they want.

Another licensing requirement that makes for a CC-type AI work is applying the principle of share-alike to the prompt settings: you could license a model such that works generated with the model must be shared with both a reference to the model and the prompt/settings used in creation (usually about a sentence of plain text).

This would not allow people to scientifically recreate exactly the same output, but it is a significant step towards identifying which source images in the data set used to train the model impacted the final product.

This “prompt sharing” is a thing AI artists are already doing, with the explicit intent of sharing insight into their work and making it easier to build on creatively; so this would not be a new invention of a license, but rather a codification of what is already the best practice for knowledge sharing.

Derivative models🔗

It is also possible to create models by merging/processing existing models instead of images.

sequenceDiagram
    Alice->>Model: Hello. Here are N models, instead of images.

The share-alike principle should apply here. CC-ish licensed models should require that any models made from it is licensed under the same license (or one more permissive) to ensure the work is shared-alike and to prevent trivial laundering.

Other interfaces, tooling🔗

There is also software that provides an interface to an existing model so people can more easily use them. These can range from anything from scratch python code to Google Colab notebooks to polished mobile apps.

There isn’t anything much novel about them, from a copyright perspective: they’re pieces of interface software, and shouldn’t have much to do with the copyright status of the models they use or the outputs they generate unless they’re actively violating an existing license.

horizontal rule

cyber Lies, Damned Lies, and Subscriptions

  • Posted in cyber

Everybody hates paying subscription fees. At this point most of us have figured out that recurring fees are miserable. Worse, they usually seem unfair and exploitative. We’re right about that much, but it’s worth sitting down and thinking through the details, because understanding the exceptions teaches us what the problem really is. And it isn’t just “paying people money means less money for me”; the problem is fundamental to what “payment” even is, and vitally important to understand.

Human Agency: Why Property is Good🔗

or, “Gio is not a marxist, or if he is he’s a very bad one”

First: individual autonomy — our agency, our independence, and our right to make our own choices about our own lives — is threatened by the current digital ecosystem. Our tools are powered by software, controlled by software, and inseparable from their software, and so the companies that control that software have a degree of control over us proportional to how much of our lives relies on software. That’s an ever-increasing share.

politics 5G's standard patents wound it

I remember seeing a whole kerfuffle about 5G around this time last year. Not the mind-control vaccine, the actual wireless technology. People (senators, mostly) were worried about national security, because Huawei (the state-controlled Chinese tech company, who is a threat, actually) was getting its 5G patents through and making its claim on the next-gen tech IP landscape. Maybe Trump even needed to seize the technology and nationalize 5G? Everybody sure had a lot to say about it, but I didn’t see a single person address the core conflict.

Format Wars🔗

Before we get to 5G, let’s go way back to VHS for a minute.

The basic idea of the “format war” is this: one company invents a format (VHS, SD cards, etc) and make a push to make their format the standard way of doing things. Everybody gets a VHS player instead of BetaMax, so there’s a market for the former but not for the latter. Now everyone uses VHS. If you’re selling video, you sell VHS tapes, and if you’re buying video, you’re buying VHS. If you invented VHS, this is great for you, because you own the concept of VHS and get to charge everyone whatever you want at every step in the process. And, since everyone uses VHS now, you’ve achieved lock-in.

Now, this creates an obvious perverse incentive. Companies like Sony are famous for writing and patenting enormous quantities of formats that never needed to exist in the first place because owning the de factor standard means you can collect rent from the entire market. That’s a powerful lure.

And that’s just talking about de facto standards. This gets even worse when you mix in formal standards setting bodies, which get together and formally declare which formats should be considered “standard” for professional and international use. If you could get your IP written into those standards, it turns your temporary development time into a reliable cash stream.

Enter SEPs🔗

“5G” is one of these standards set by standard setting bodies, and it’s a standard packed with proprietary technology. The most important slice of those is called SEPs, or “Standard Essential Patents.” These are the Patents that are Essential to (implementing) the Standard. In other words, these technologies are core and inextricable to 5G itself. This figure represents only the SEPs:

fandom Trouble a-brewin' at Redbubble

  • Posted in fandom

Homestuck is once again lit up over fan merch. Homestuck and fan merch have a long and troubled history, but this latest incident is between artists, Redbubble, and Viz media. Here are my thoughts on that!

In late May 2021, artists who sold Homestuck merch on Redbubble got this email:

Dear [name],

Thank you for submitting your fan art for Homestuck and/or Hiveswap as part of Redbubble’s Fan Art Partner Program.

At this time, our partnership with the rights holder VIZ Media has come to an end. When a partnership expires, we are required to remove officially approved artworks from the marketplace. This means that your Homestuck and/or Hiveswap designs will be removed from Redbubble soon.

Here are a couple of things to keep in mind:

  • It is important to know that licensors do not allow previously approved designs once sold on Redbubble to be sold on any other platform, even after the program ends.
  • Because this removal is not in response to a complaint, your account will not be negatively impacted.

Partnerships come and go, but don’t worry. We’re looking forward to partnering with more awesome brands in the future.

Check out our Current Brand Partnerships list to see all the properties that are actively accepting submissions. For additional information, we recommend checking out the Fan Art Partner Program FAQ.

Thank you, Redbubble

This hit a lot of people, and hit them hard:

Rut-roh!

Unfortunately for Twitter and brevity this is actually the intersection of a couple different complicated issues, which I’ll try to summarize here.

Redbubble forcing predatory licensing on people🔗

Now, copyright law sucks for fanartists, but that doesn’t explain what happened here.