DCT
6:24-cv-00245
Cellular South Inc v. Google LLC
I. Executive Summary and Procedural Information
- Parties & Counsel:- Plaintiff: Cellular South, Inc. (d/b/a C Spire) (Mississippi)
- Defendant: Google, LLC (Delaware)
- Plaintiff’s Counsel: Holland & Knight LLP; The Dacus Firm, P.C.
 
- Case Identification: 6:24-cv-00245, W.D. Tex., 05/09/2024
- Venue Allegations: Plaintiff alleges venue is proper because Google maintains an established place of business in the district, specifically an office in Austin, Texas, and has committed acts of infringement in the district.
- Core Dispute: Plaintiff alleges that Defendant’s Google Cloud Video Intelligence and Video AI platforms infringe three patents related to organizing and analyzing unstructured data within video content.
- Technical Context: The technology at issue involves using machine learning and parallel processing to automatically analyze video content, extracting and cross-referencing information from both image frames and audio tracks to create searchable, contextual metadata.
- Key Procedural History: The complaint alleges that Plaintiff's subsidiary, Vū Digital, presented its "patent pending" V2D technology to corporate partners, including Google, at a SPROCKIT Sync conference on June 18, 2015. It also alleges Google identified the ’954 patent as potential prior art during the prosecution of its own patent application in November 2022, suggesting pre-suit knowledge of the patent family.
Case Timeline
| Date | Event | 
|---|---|
| 2013-08-15 | Earliest Priority Date for '954, ’972, and ’853 Patents | 
| 2015-05 | Plaintiff's V2D product debuts | 
| 2015-06-18 | Plaintiff's subsidiary allegedly presents "patent pending" tech to Google | 
| 2018-04-10 | U.S. Patent No. 9,940,972 Issues | 
| 2019-02-26 | U.S. Patent No. 10,218,954 Issues | 
| 2021-09-21 | U.S. Patent No. 11,126,853 Issues | 
| 2022-11-02 | Google allegedly identifies '954 Patent in an IDS for its own application | 
| 2024-05-09 | Complaint Filed | 
II. Technology and Patent(s)-in-Suit Analysis
U.S. Patent No. 10,218,954 - "Video to Data"
Issued February 26, 2019
The Invention Explained
- Problem Addressed: The patent describes the difficulty of accurately identifying objects in video frames, which can lead to "false positives" (Compl. ¶32; ’954 Patent, col. 5:3-7). Standalone image analysis lacks the context provided by other parts of the video, such as the audio track or the sequence of preceding and subsequent frames (Compl. ¶32; ’954 Patent, col. 5:20-25).
- The Patented Solution: The invention proposes a method that generates both audio and image files from a video and processes them in parallel (Compl. ¶37; ’954 Patent, Abstract). It uses context from both the audio (e.g., an announcer's commentary) and surrounding image frames (e.g., identifying a horse race) to assign a probability of accuracy to an identified object, thereby reducing errors (Compl. ¶32; ’954 Patent, col. 5:7-16). This cross-referenced data is then used to generate a "content-rich video" (Compl. ¶37; ’954 Patent, col. 14:19-21).
- Technical Importance: This approach aimed to make video content as searchable as text by systematically deriving contextual meaning from the relationship between visual and audio data streams (Compl. ¶19).
Key Claims at a Glance
- The complaint asserts independent claims 1 and 13 (Compl. ¶36).
- Independent Claim 1 recites a method with the essential elements of:- generating audio files and image files from a video;
- distributing and processing these files in parallel;
- converting audio files to text;
- identifying an object in the image files;
- determining a contextual topic from the image files;
- assigning a probability of accuracy to the object based on the contextual topic;
- converting the image files to video data comprising the object, probability, and topic;
- cross-referencing the text and video data to determine contextual topics; and
- generating a content-rich video based on the results.
 
- The complaint reserves the right to assert corresponding dependent claims (Compl. ¶36).
U.S. Patent No. 9,940,972 - "Video to data"
Issued April 10, 2018
The Invention Explained
- Problem Addressed: As with the '954 Patent, the '972 Patent addresses the challenge of extracting meaningful metadata from video, but focuses more specifically on how to combine information derived from different data streams (Compl. ¶74).
- The Patented Solution: The invention describes a method where topics generated from image analysis and topics extracted from audio analysis can be combined and cross-referenced (’972 Patent, col. 5:46-49). This combination allows for the generation of "topical meta-data" that reflects a more holistic understanding of the video's content (Compl. ¶78). This metadata is then added back to the video, and a new text description, image, or animation can be generated and placed into the video (’972 Patent, col. 6:23-29).
- Technical Importance: The invention provided a method for creating a unified, semantically rich metadata layer by explicitly combining information derived separately from the video's visual and audio components, allowing for more complete summaries (Compl. ¶75; ’972 Patent, col. 5:49-54).
Key Claims at a Glance
- The complaint asserts independent claims 1 and 17 (Compl. ¶77).
- Independent Claim 1 recites a method with the essential elements of:- generating audio and image files from a video;
- distributing and processing image files in parallel to extract objects;
- processing audio files and converting them to text;
- converting image files to video data;
- generating a "topical meta-data" by deriving semantic information from both the identified objects and the audio files;
- adding the topical meta-data to the video;
- cross-referencing the text and video data based on the topical meta-data to determine topics; and
- generating and placing a text, image, or animation into the video.
 
- The complaint reserves the right to assert corresponding dependent claims (Compl. ¶77).
U.S. Patent No. 11,126,853 - "Video to Data"
Issued September 21, 2021
The Invention Explained
- The '853 Patent describes a system for improving object recognition, particularly in the presence of visual "noise" or obstructions, by using an adjustable image detector and an object recognizer. The recognizer compares an object's image to a "fractal," which is a representation based on landmarks associated with the object, and is configured to update that fractal with the new image, suggesting a self-improving recognition model (Compl. ¶117, ¶119-120; '853 Patent, claim 1).
Key Claims and Allegations
- Asserted Claims: The complaint asserts independent claim 1 (Compl. ¶122).
- Accused Features: The complaint alleges that Google's Video Intelligence, which uses the Cloud Vision API as an "image classifier" or "image detector," infringes the '853 Patent. It further alleges that Google's system, which is "constantly" trained and updated using data from millions of videos, meets the "object recognizer configured to... update the fractal with the image" limitation (Compl. ¶134-138). The complaint points to Google's ability to detect both "significant and less-prominent objects" as evidence of the claimed adjustable detector (Compl. ¶135, ¶62).
III. The Accused Instrumentality
Product Identification
- Google's Cloud Video Intelligence platform, also referred to as Video AI, including its API and the associated Vertex AI for AutoML platform (collectively, the "Accused Product") (Compl. ¶39, ¶40).
Functionality and Market Context
- The Accused Product is a cloud-based service that uses machine learning models to analyze video files. It can automatically recognize objects, places, actions, and text within video; transcribe speech; and generate labels or tags for video content (Compl. ¶39, ¶51).
- The platform offers both pre-trained models (Video Intelligence API) for common use cases and the ability for customers to train custom models (Vertex AI for AutoML) (Compl. ¶41). A screenshot from Google's marketing materials describes these as "Two ways to make your media more discoverable and valuable" (Compl. ¶41, p. 12).
- The complaint alleges the Accused Product is used for applications such as content moderation, creating media archives, building recommendation engines, and enabling contextual advertising (Compl. ¶40, ¶42). A testimonial from CBS Interactive included in the complaint states, "Video Intelligence allows CBS Interactive to plug into our existing video encoding framework to generate video metadata" (Compl. ¶42, p. 14).
IV. Analysis of Infringement Allegations
Infringement Allegations: U.S. Patent No. 10,218,954
| Claim Element (from Independent Claim 1) | Alleged Infringing Functionality | Complaint Citation | Patent Citation | 
|---|---|---|---|
| generating audio files and image files from the video | The Accused Product decodes an uploaded video into its constituent "audio streams, video streams, subtitles and all of the metadata." | ¶44 | col. 3:10-12 | 
| distributing the audio files and the image files across a plurality of processors and processing the audio files and the image files in parallel | The Accused Product allegedly executes on the Google Cloud Platform ("GCP"), which is described as a distributed network of physical assets in data centers around the world. A diagram in the complaint illustrates GCP's distributed architecture of regions and zones. | ¶45, p. 17 | col. 3:25-28 | 
| converting audio files associated with the video to text | The Accused Product provides a "SPEECH_TRANSCRIPTION" feature to extract and convert audio data into text. | ¶47 | col. 7:1-3 | 
| identifying an object in the image files | The system applies an "image classifier" to identify objects in frames and create annotations for "entities" within the video. | ¶49-50 | col. 3:28-32 | 
| determining a contextual topic from the image files | The Accused Product's "LABEL_DETECTION" feature applies tags or labels (e.g., "train," "transportation") to identify entities, which the complaint equates with determining a contextual topic. | ¶51 | col. 5:7-16 | 
| assigning a probability of accuracy to the identified object based on the contextual topic | Google's documentation and presentations allegedly state that the system provides a confidence or accuracy rating for identified objects, such as an "elephant with 97% accuracy." A screenshot shows frame-level entities with associated percentages. | ¶54, p. 23 | col. 5:57-63 | 
| converting the image files associated with the video to video data, wherein the video data comprises the object, the probability, and the contextual topic | The Accused Product allegedly detects entities, calculates a confidence value, and assigns a contextual label, which are elements that can be retrieved via API calls like "LabelSegment." | ¶56 | col. 14:10-15 | 
| cross-referencing the text and the video data with the video to determine contextual topics | The system is alleged to use an "aggregating video classifier" that combines outputs from image, audio, and text analysis to determine a contextual topic like "a documentary about animals in Africa." | ¶58 | col. 14:15-18 | 
| generating a contextual text, an image, or an animation based on the determined contextual topics | The complaint alleges Google's platform can be used to generate highlight reels, create video descriptions for searching, or automatically generate thumbnail images. | ¶61-62 | col. 14:18-21 | 
| generating a content-rich video based on the generated text, image, or animation | This is allegedly performed when the system creates highlight reels by stitching together moments identified through analysis of the contextual topics. | ¶61 | col. 14:19-21 | 
- Identified Points of Contention:- Scope Questions: A central question may be whether Google's "LABEL_DETECTION" feature, which applies tags like "train" or "river," performs the function of "determining a contextual topic" as that term is used in the patent. The patent provides examples like "football game" or "horse race," which suggests a higher-level thematic understanding than a simple object label (’954 Patent, col. 5:7-16).
- Technical Questions: The complaint alleges infringement by combining discrete features described in various Google marketing materials and technical presentations. A key question for the court will be whether the Accused Product actually performs the specific, ordered sequence of steps in Claim 1, particularly the step of creating a unified "video data" that comprises the object, probability, and topic before cross-referencing it with the text.
 
Infringement Allegations: U.S. Patent No. 9,940,972
| Claim Element (from Independent Claim 1) | Alleged Infringing Functionality | Complaint Citation | Patent Citation | 
|---|---|---|---|
| generating audio files and image files from the video | The Accused Product decodes a video into audio and video streams. | ¶85 | col. 5:28-29 | 
| distributing the image files across a plurality of processors and processing the image files in parallel, wherein processing the image files comprises extracting one or more objects and identifying the one or more objects | The Accused Product allegedly runs on the distributed Google Cloud Platform to identify and label entities at the frame, shot, or video level. | ¶86, ¶91 | col. 5:30-34 | 
| converting audio files associated with the video to text | The system's "SPEECH_TRANSCRIPTION" feature converts audio to text. | ¶88-89 | col. 5:37-38 | 
| generating a topical meta-data that describes content of the video by deriving semantic information from the identification of the one or more objects and semantic information from the audio files | The complaint alleges Google's "aggregating video classifier" combines outputs from analyzed image data, audio, and text to determine a contextual topic, which it equates to generating "topical meta-data." | ¶95 | col. 5:44-49 | 
| adding the topical meta-data to the video | This is allegedly accomplished by associating the generated "entities," "labels," and "tags" with the video file, which can then be accessed via API calls. | ¶99 | col. 6:1-2 | 
| cross-referencing the text and the video data based on the generated topical meta-data to determine topics | The complaint again points to the "aggregating video classifier" that combines outputs from multiple data streams to determine a contextual topic for a given video. | ¶100 | col. 6:2-4 | 
| generating video text based on the cross-referencing, wherein the video text describes content of the video | The complaint alleges the Accused Product generates video descriptions or summaries based on the extracted semantic information, for example to create highlight reels. A diagram illustrates a workflow for building media portals with "video summaries." | ¶102-103, p. 47 | col. 6:4-6 | 
| generating a text, image, or animation based on the video text; and placing the text, image, or animation in the video | The complaint alleges this is performed when the system generates and places captions on a video or creates thumbnails. | ¶104, ¶107 | col. 6:23-29 | 
- Identified Points of Contention:- Scope Questions: The infringement theory hinges on whether Google's combination of analyzed data constitutes "generating a topical meta-data" by deriving semantic information from both audio and image sources, as specifically claimed. The defense may argue that its system generates separate data streams that are merely presented together, rather than creating a single, new "topical meta-data" element as the claim requires.
- Technical Questions: What evidence does the complaint provide that the Accused Product "add[s] the topical meta-data to the video" and then uses that specific added metadata as the basis for the subsequent cross-referencing step? The claim implies a specific data flow that may be a point of dispute.
 
V. Key Claim Terms for Construction
U.S. Patent No. 10,218,954
- The Term: "contextual topic"
- Context and Importance: This term is central to the patent's claimed improvement over prior art object identification. The infringement case depends on whether Google's "labels" (e.g., "river," "train") qualify as a "contextual topic." Practitioners may focus on this term because its definition will determine if Google's granular object tagging meets the claim's requirement for a higher-level thematic determination.
- Intrinsic Evidence for Interpretation:- Evidence for a Broader Interpretation: The claim language recites "determining a contextual topic from the image files," which could be read to mean any topic, including a simple object name, derived from the image.
- Evidence for a Narrower Interpretation: The specification provides examples of contextual topics such as "a football game" or "a horse race," which are derived from analyzing multiple objects and their relationships, not just identifying a single object. This suggests a more thematic or event-based definition than a simple label (’954 Patent, col. 5:9-16).
 
U.S. Patent No. 9,940,972
- The Term: "topical meta-data"
- Context and Importance: This term defines the core output of the claimed combination step. The dispute will likely focus on whether the output of Google's "aggregating video classifier" is a new data element that meets this definition. Practitioners may focus on this term because the claim requires this specific data element to be generated, added to the video, and then used as the basis for further cross-referencing.
- Intrinsic Evidence for Interpretation:- Evidence for a Broader Interpretation: The term is not explicitly defined, which might support an argument that any combination of topic information from audio and video sources qualifies.
- Evidence for a Narrower Interpretation: The claim requires this metadata to be generated by "deriving semantic information from the identification of the one or more objects and semantic information from the audio files." This "and" suggests the "topical meta-data" must be a unified element created from both sources, not just a collection of separate tags from each source (’972 Patent, col. 5:44-49). The specification reinforces this, stating "the topics generated from an image...and the topics extracted from audio can be combined" (’972 Patent, col. 5:46-47).
 
VI. Other Allegations
- Indirect Infringement: The complaint does not contain a separate count for indirect infringement. However, it alleges that Google "directs and controls each relevant aspect of the accused technology" and provides documentation and APIs that enable and encourage customers to use the technology in an infringing manner, which could potentially support future indirect infringement claims (Compl. ¶3, ¶40).
- Willful Infringement: Willfulness is alleged for all three patents. The claims are based on alleged pre-suit knowledge stemming from Google's alleged participation in the 2015 SPROCKIT conference where the technology was presented, and from Google's own Information Disclosure Statement filed in 2022 that cited the '954 patent (Compl. ¶66-67, ¶110-111, ¶143).
VII. Analyst’s Conclusion: Key Questions for the Case
- A core issue will be one of technical implementation: Do Google's cloud-based AI services, which are presented as a flexible suite of tools, actually perform the specific, ordered steps recited in the method claims? The case may turn on evidence of the precise data flow within Google's black-box systems versus the sequential process described in the patents.
- A second key question will be one of definitional scope: Can the patent term "fractal," a precise mathematical concept used in the '853 patent, be construed to read on Google's use of constantly evolving machine-learning models for object recognition? The plaintiff's ability to equate Google's proprietary AI training methods with the specific "fractal" structure claimed will be a central point of contention.
- Finally, the dispute raises a question of semantic interpretation: Do the simple "labels" and "tags" generated by Google's Video AI (e.g., "elephant," "river") meet the patents' requirements for a "contextual topic" ('954 patent) or "topical meta-data" derived from both audio and video sources ('972 patent), which the specifications suggest are higher-level, thematic descriptors?