Trademark

Hayley Paige Turns the Final Page

The long-running dispute between wedding dress designer Hayley Paige Gutman and her former employer JLM Couture over ownership of the social media accounts she created and that bear her name is at last over.

On May 8, the District Court revised its earlier preliminary injunction and restored control of the @misshayleypaige Instagram, Pinterest and TikTok accounts to Gutman. In its decision, the Court held JLM couldn’t establish that the social media accounts belonged to JLM at their inception or that Gutman had transferred ownership to JLM. 

In case you’re new to my analysis of the very public case (at least to Gutman’s 1 million-plus followers), I originally wrote about it here and, more recently, here

How did the District Court reach a conclusion that is 180 degrees from its earlier decisions? 

Well, in March 2024 the Second Circuit disagreed with the six-factor test the District Court created to determine ownership of the social media accounts and remanded the case back to the District Court with instructions to look at the accounts as normal property rather than through novel tests (a decision I heartily agreed with). After that, the District Court requested and received additional briefing on whether JLM could show that it either owned the accounts at the time they were created or that ownership was subsequently transferred to JLM, and that Gutman only had access to the social media accounts because she was a JLM employee.

In its analysis, the Court found:

  • Gutman opened the Instagram account at the recommendation of a friend;
  • She was at least partially motivated to create the accounts for personal use even if she saw them as potentially useful in promoting products manufactured by JLM; 
  • The Instagram was initially linked to her personal Facebook account;
  • The Instagram account wasn’t linked to the JLM Facebook account until four years later; and 
  • The first five Instagram posts were clearly personal, as were early pins on Pinterest 

According to the Court, taken together, these facts showed that Gutman created the accounts for personal use and not solely as a JLM employee, even if she did later use them to promote JLM products, allowed JLM employees access to them, and included links to JLM accounts. 

Thus, since JLM couldn’t establish a valid claim to ownership of the social media accounts when they were created, it would have to show that the accounts were transferred to it. It was unable to do this for several reasons including the fact that Instagram’s terms of service prevent a user from transferring an account.

There were a few other interesting factors in the mix. For one, while Gutman’s contract with JLM included licensing her name to JLM for commercial purposes, she retained the right to use her name for non-commercial purposes. Thus, the Court found her use of @misshayleypaige as a handle on social media was not on its own a basis for finding that her social media accounts belong to JLM. And, as the Second Circuit found, the social media accounts themselves do not qualify as works for hire.

Gutman had also signed an employment manual with JLM in which she agreed that she would use her time during her workday only for employment-related activities; that her use of the Internet during that time was for job-related activities; and all intellectual properties she created belonged to JLM. The order modifying the preliminary injunction noted that even if Gutman’s posting on social media during working hours was a violation of JLM’s employment manual, it doesn’t mean that, despite JLM’s arguments, it was tantamount to her transferring her social media accounts to JLM. It does suggest, however, that JLM has a legitimate claim to the content Gutman posted on her social media accounts.

What’s really interesting here is that, in the end, the final decision in the District Court is opposite to the results reached by just about every other court that has looked at usage to determine ownership of a social media account. Will the Gutman decision reverberate throughout the growing number of legal disputes in this area, and give pause to companies who rely heavily on consumer relationships through the Instagram posts of highly visible employees and brand ambassadors?

What’s next? Well, not much. Shortly after the most recent decision from the District Court, Gutman and JLM resolved all matters in a settlement agreement that included Gutman paying JLM (well, its bankruptcy estate) $263,000. In exchange, Gutman was released from her non-compete agreement which would have continued for another 18 months or so.

 

Chanel Goes Coco Over Unauthorized Sales, Hashtags

In March 2018, iconic fashion house Chanel sued What Goes Around Comes Around (“WGACA”), a reseller of luxury goods à la Poshmark and The RealReal (which Chanel has also sued). These retailers are essentially online thrift stores (WGACA is also brick-and-mortar), solely trafficking in high-end designer goods and apparel instead of ratty Wranglers and stained JCPenney tops. 

WGACA, Poshmark, The RealReal, and others of their ilk tout how recycling luxury items is good for the environment (not to mention their bottom lines); Chanel, however, finds their practices less than noble. In its lawsuit, Chanel accused WGACA of trademark infringement and false advertising by selling unauthorized Chanel products and using the Chanel trademark too prominently in its marketing. I wanted to highlight this case because it’s an important lesson in what is and is not permitted when it comes to using someone else’s trademarks and other materials associated with a brand.

Full disclosure: Another reason I find this case interesting is that as the quality of mass retailers gets worse and worse every year, the only way for me to find well-made goods at a non-astronomical price is through the resale market. So yeah, I’m a little personally involved here.

Chanel, Inc. v. WGACA, LLC, went to trial in January of this year. On February 6, the jury returned a verdict for Chanel and awarded them $4 million in statutory damages for willful trademark infringement. How did they reach this conclusion? 

Let’s start with some background. Under the first sale doctrine, once a genuine product is sold, the person who purchases it is free to resell it without risk of liability to the brand owner. Likewise, it’s totally fine to use someone else’s brand or trademark to accurately describe the pre-owned item in sales materials. For example, I can snap a photo of a pair of Nike Air Force 1s or a Birkin bag taking up space in my closet, list it for sale on eBay or Facebook Marketplace and use the brand name in my written description along with pics of the brands’ trademarks (assuming the item I’m listing is genuine) — provided I use the marks or brand names only to the extent necessary to describe what I’m selling, and I don’t do anything that might suggest that I’m affiliated with Nike or Hermès or that either company is endorsing my resale of the items. 

While there was no question that many (although not all) of the goods WGACA offered were genuine, WGACA ran into trouble with Chanel because there was significant evidence that WGACA used Chanel’s marks too prominently and too often. For example, on social media, it hashtagged posts with #WGACACHANEL. Elsewhere, WGACA featured a Chanel mark more prominently than its own brand mark. Moreover, WGACA’s website and other communications included the statement “WGACA CHANEL – 100% Authenticity Guaranteed.” The jury appears to have viewed this as WGACA suggesting it was endorsed by or had a relationship with Chanel, and on this basis, ruled for Chanel. 

WGACA also sold items that, according to Chanel, were never approved for retail, including handbags with voided or pirated serial numbers as well as decorative items Chanel lent to retailers. On this issue the District Court granted summary judgment in favor of Chanel, finding these items were never authorized for sale by anyone, much less WGACA. 

What lessons should resellers take away from this? For starters, there’s plenty of room to truthfully advertise the resale of authentic luxury goods. However, it’s important to have specific processes in place to vet any advertising to make sure it doesn’t suggest an affiliation with the brand where there isn’t one. In a similar vein, resellers need to develop programs that ensure the goods they offer for sale are authentic and were originally sold by the brand or, if that’s not always possible, to accurately communicate with consumers. 

But the case isn’t closed yet. It remains before the District Court on Chanel’s request that WGACA be prevented from, among other things, using Chanel’s marks to promote WGACA’s business; including the word Chanel in any hashtags; and using in its advertising any Chanel-branded items other than items actually for sale by WGACA. 

I’m curious to see what happens at the preliminary injunction hearing. Here, I’m particularly interested in what the Court has to say about the use of hashtags on social media because this seems like an area where, to date, most courts have assumed that a hashtag using a brand name is infringing or gives rise to liability. Perhaps this Court will give us a more nuanced analysis. I hope so, since I believe that resellers like WGACA and The RealReal will continue to thrive as consumers look to both save money and purchase goods with a smaller environmental impact which, as mentioned above, is a key element in these resellers’ marketing. Of course, one could question whether their “green” claims constitute false advertising, but that’s a question for another day.

Let’s Talk About Trademarks (And AI)

I’ve posted quite a bit about the growing legal battles involving AI companies, copyright infringement, and the right of publicity. These are still early days in the evolution of AI so it’s hard to envision all the ways the technology will develop and be utilized, but I predict AI is going to come up against even more existing intellectual property laws — specifically, trademark law.

For example, in its lawsuit against Open AI and others (which I wrote about here), the New York Times Company alleged the Defendants engaged in trademark dilution. To take a step back, trademark dilution happens when someone uses a “famous” trademark (think Nike, McDonalds, UPS, etc.) without permission, in a way that weakens or otherwise harms the reputation of the mark’s owner. This could happen when an AI platform, in response to a user query, delivers flat-out wrong or offensive content and attributes it to a famous brand such as the New York Times. Thus, according to the Times’ complaint, when asked “what the Times said are ‘the 15 most heart-healthy foods to eat,’” Bing Chat (a Microsoft AI product) responded with, among other things, “red wine (in moderation).” However, the actual Times article on the subject “did not provide a list of heart-healthy foods and did not even mention 12 of the 15 foods identified by Bing Chat (including red wine).” Who knows where Bing got its info from, but if the misinformation and misattribution causes people to think less of the “newspaper of record,” that could be construed as trademark dilution.  

There are, however, potential pitfalls for brands who want to use trademark dilution to push back against AI platforms. It’s difficult to discover, expensive to pursue and there can be a lot of ambiguity about whether a brand is “famous” and able to be significantly harmed by trademark dilution. In the New York Times case, the media giant has the resources to police the Internet and to file suits; nor should there be any dispute that is a “famous” brand with a reputation that is vitally important. But smaller companies may not have the resources to search for situations where AI platforms incorrectly attribute information, or have a platform visible enough to meaningfully correct the record. Plus, calculating the brand damage from AI “hallucinations” will be very difficult and costly.  Also, this area of the law does nothing for brands that aren’t “famous.” 

Another area where trademark law and AI seem destined to face off is under the sections of the Lanham Act — the Federal trademark law — that allows celebrities to sue for non-consensual use of their persona in a way that leads to consumer confusion, or others to sue for false advertising that influences consumer purchasing decisions. AI makes it pretty easy to manipulate a celebrity’s (or anyone’s) image or video to do and say whatever a user wants, which opens up all sorts of troublesome trademark possibilities.

Again, there are a couple of serious limitations here. For starters, the false endorsement prong likely only applies to celebrities or others who are well-known and does little to protect the rest of us. Perhaps more important (and terrifying), it seems likely that there will be significant issues in applying the Lanham Act’s provisions on false advertising in the context of deepfakes in political campaigns — like, for example, the recent robocall in advance of the New Hampshire primary that sounded like it was from President Biden. To avoid problems with the First Amendment, the Lanham Act is limited to commercial speech and thus will be largely useless for dealing with this type of AI abuse.

One other potentially interesting (and creepy) area where AI and trademark law might intersect is when it comes to humans making purchasing decisions through an AI interface. For example, a user tells a chatbot to order a case of “ShieldSafe disinfecting wipes,” but what shows up on their porch is a case of “ShieldPro disinfecting wipes” (hat tip to ChatGPT for suggesting these fictional names). While the mistake of a few letters might mean nothing to an algorithm (or even to a consumer who just wants to clean a toilet), it’s certainly going to anger a ShieldSafe Corp. that wants to prevent copycat companies from stealing their customers (and keep their business from going down that aforementioned toilet). 

When the Test is a Fail

The Second Circuit recently issued a(nother) decision in the dispute between bridalwear designer Hayley Paige Gutman and her former employer, JLM Couture, Inc. over ownership of Instagram and Pinterest accounts Gutman created while employed by JLM. You can find background on this case here

The Second Circuit reversed the District Court’s 2022 decision, which held JLM owned the accounts. More notably, the Second Circuit rejected the lower court’s six factor test considering how the account describes itself; whether the account was promoted on the employer entity’s advertisements or publicity materials and linked to other internet platforms of the entity; whether it promoted the business; and whether employees of the entity (other than the account creator) had access to and managed the account. 

In reversing the lower court, the Second Circuit held social media accounts “should be treated in the first instance like any other form of property,” and, in figuring out who currently owns one, courts should look to who owned it when it was created and whether there is any evidence the account was ever transferred to someone else. “[T]he law has long accommodated new technologies within existing legal frameworks,” the Circuit wrote. Translation: “Enough with the new tests already. We have plenty.” 

Overall, the Second Circuit’s conclusion lines up with what I suggested in a New York Law Journal article last month: Who Owns a Social Media Account? It’s Pretty Simple, Really. The article is paywalled, but the gist is this: Courts should stop coming up with new tests to determine whether a social media account belongs to a business or an individual associated with the business and, instead, look to existing and well-established legal frameworks to determine ownership. 

Hopefully, this is what will happen when the District Court takes up Gutman’s case again. Here, the Second Circuit sent the case back to the lower court with a note that the ownership of the social accounts may turn, at least in part, on the terms of service of the relevant social media platforms and, specifically, does “ownership” of a social media account include the right to transfer the account to another. The Second Circuit also suggested that the District Court might want to separate the ownership of content posted on the accounts from the ownership of the accounts themselves, noting that rights to the accounts and rights to the accounts’ content may or may not be the same.   

And on it goes. I’ll be paying particular attention to what the District Court says about the role of the terms of service for social media accounts. As we all know, social media companies change their terms of service — a lot. Does this approach give an outsized role to the terms of service even though the litigants in cases over ownership of social media accounts have no input into the terms of those agreements? Second, how will courts factor in changes to terms of service that parties may or may not be aware of, particularly as at least one of the parties to a dispute over ownership of a social media account probably never agreed to the terms of service? I suspect these issues will mean the District Court downplays the significance of the terms of service and instead looks at doctrines governing the ownership of other intangible property. Because tests that are well established… tend not to fail.

Sign of the Times: The Battle Against AI Goes Big

I closed out 2023 by writing about one lawsuit over AI and copyright and we’re starting 2024 the same way. In that last post, I focused on some of the issues I expect to come up this year in lawsuits against generative AI companies, as exemplified in a suit filed by the Authors Guild and some prominent novelists against OpenAI (the company behind ChatGPT). Now, the New York Times Company has joined the fray, filing suit late in December against Microsoft and several OpenAI affiliates. It’s a big milestone: The Times Company is the first major U.S. media organization to sue these tech behemoths for copyright infringement. 

As always, at the heart of the matter is how AI works: Companies like OpenAI ingest existing text databases, which are often copyrighted, and write algorithms (called large language models, or LLMs) that detect patterns in the material so that they can then imitate it to create new content in response to user prompts.

The Times Company’s complaint, which was filed in the Southern District of New York on December 27, 2023, alleges that by using New York Times content to train its algorithms, the defendants directly infringed on the New York Times copyright. It further alleges that the defendants engaged in contributory copyright infringement and that Microsoft engaged in vicarious copyright infringement. (In short, contributory copyright infringement is when a defendant was aware of infringing activity and induced or contributed to that activity; vicarious copyright infringement is when a defendant could have prevented — but didn’t — a direct infringer from acting, and financially benefits from the infringing activity.) Finally, the complaint alleges that the defendants violated the Digital Millennium Copyright Act by removing copyright management information included in the New York Times materials, and accuses the defendants of engaging in unfair competition and trademark dilution. 

The defendants, as always, are expected to claim they’re protected under “fair use” because their unlicensed use of copyrighted content to train their algorithms is transformative. 

What all this means is that while 2023 was the year that generative AI exploded into the public’s consciousness, 2024 (and beyond) will be when we find out what federal courts think of the underlying processes fueling this latest data revolution.

I’ve read the New York Times’ complaint (so you don’t have to) and here are some takeaways:

  • The Times Company tried (unsuccessfully) to negotiate with OpenAI and Microsoft (a major investor in OpenAI) but were unable to reach an agreement that would “ensure [The Times] received fair value for the use of its content.” This likely hurts the defendants’ claims of fair use. 
  • As in the other lawsuits against OpenAI and similar companies, there’s an input problem and an output problem. The input problem comes from the AI companies ingesting huge amounts of copyrighted data from the web. The output problem comes from the algorithms trained on the data spitting out material that is identical (or nearly identical) to what they ingested. In these situations, I think it’s going to be rough going for the AI companies’ fair use claim. However, they have a better fair use argument where the AI models create content “in the style of” something else.
  • The Times Company’s case against Microsoft comes, in part, from the fact that Microsoft is alleged to have “created and operated bespoke computing systems to execute the mass copyright infringement . . .” described in the complaint.
  • OpenAI allegedly favored “high-quality content, including content from the Times” in training its LLMs.
  • When prompted, ChatGPT can regurgitate large portions of the Times’ journalism nearly verbatim. Here’s an example taken from the complaint showing the output of ChatGPT on the left in response to “minimal prompting,” and the original piece from the New York Times on the right. (The differences are in black.)

Excerpt from The New York Times Company's Complaint

  • According to the New York Times this content, easily accessible for free through OpenAI, would normally only be available behind their paywall. The complaint also contains similar examples from Bing Chat (a Microsoft product) that go far beyond what you would get in a normal search using Bing. (In response, OpenAI says that this kind of wholesale reproduction is rare and is prohibited by its terms of service. I presume that OpenAI has since fixed this issue, but that doesn’t absolve OpenAI of liability.)
  • Because OpenAI keeps the design and training of its GPT algorithms secret, the confidentiality order here will be intense because of the secrecy around how OpenAI created its LLMs.
  • While the New York Times Company can afford to fight this battle, many smaller news organizations lack the resources to do the same. In the complaint, the Times Company warns of the potential harm to society of AI-generated “news,” including its devastating effect on local journalism which, if the past is any indication, will be bad for all of us

Stay tuned. OpenAI and Microsoft should file their response, which I expect will be a motion to dismiss, in late-February or so. When I get those, I’ll see you back here.