AI/ML

TWLV-I: Analysis and Insights from Holistic Evaluation on Video Foundation Models

Twelve Labs introduces a robust evaluation framework for video understanding, emphasizing both appearance and motion analysis.

Lucas Lee, Kilian Baek, James Le
Embeddings
Foundation models
Multimodal AI
Research
Video understanding
Pegasus-1 Open Beta: Setting New Standards in Video-Language Modeling

Our video-language foundation model, Pegasus-1. gets an upgrade!

Minjoon Seo, James Le
Foundation models
Generate API
Generative AI
Multimodal AI
Video understanding
Introducing Marengo-2.6: A New State-of-the-Art Video Foundation Model for Any-to-Any Search

This blog post introduces Marengo-2.6, a new state-of-the-art multimodal embedding model capable of performing any-to-any search tasks.

Aiden Lee, James Le
Foundation models
Classification API
Multimodal AI
Search API
Video understanding
Introducing Video-To-Text and Pegasus-1 (80B)

This article introduces the suite of video-to-text APIs powered by our latest video-language foundation model, Pegasus-1.

Aiden Lee
Aiden Lee, Jae Lee
Multimodal AI
Foundation models
Generate API
Generative AI
The Multimodal Evolution of Vector Embeddings

This post will give a brief definition of embeddings, walk through various unimodal embeddings, explore multimodal video embeddings, and glance at embeddings in production.

James
James Le
Transformers
Multimodal AI
Video understanding
Embeddings
What Is Multimodal AI?

Applications, Principles, and Core Research Challenges in Multimodal AI

James Le
Multimodal AI
Foundation models
Transformers
Applications
The Past, Present, and Future of Video Understanding Applications

A review of how far video understanding research has come, what potential remains untapped, and where it is headed in the future

James Le
James Le
Video understanding
Transformers
Multimodal AI
Foundation models
What makes Foundation Models special?

Capabilities and Applications of Foundation Models in Layman Terms

James Le
James Le
Foundation models
Applications
Transfer Learning
Fine-Tuning
Multimodal AI
Foundation models are going multimodal

A primer on foundation models: what they are, how they've evolved, and where they're going.

James Le
James Le
Multimodal AI
Video understanding
Foundation models
Transformers
Scaling laws

Charles Frye

Multimodality Unblocks Robots

Brian Correa

How we search and edit media with AI

Jul 07, 2023 1:30-2:30 PST
Watch now

James Le

01 An overview of foundation models and what distinguishes them from conventional approaches

02 Evolution of large language models

03 Multimodal foundation models: Vision-language and Video foundation models

Jun 30, 2023 1:30-2:30 PST
Watch now

Travis Couture

Onboarding crash course for Twelve Labs Video Understanding and Search

01 What is Twelve Labs?

02 How to use the Twelve Labs Playground

03 How to go from Playground to API: Extended functionalities

Jun 23, 2023 10:30-11:30 PST
Watch now

Vishakha Gupta

Are Vector Databases Enough for Visual Data Use Cases?

Pranay Pasula

Multimodal Learning for Learning: Perspectives and Applications

Harpreet Sahota

Lessons Learned from Building YOLO-NAS

Jun 16, 2023 1:30-2:30 PST
Watch now

Nicket Uttarwar

How many AI's does it take to run a media company?

Chris Kennedy

GAIB: Making AI chat a multimedia experience exploration

Japinder Sandhu

AI in Filmmaking: Elevating Storytelling with Computer Vision

Jun 09, 2023 1:30-2:30 PST
Watch now

Sunny Nguyen

Semantic Video Search: How to craft the perfet natural language query

Ankit Khare

Combining logical operators to create complex queries

Jun 02, 2023 1:30-2:30 PST
Watch now

Ryan Monteith

Multimodal Search in Video Editing

Fabio Caironi

Breaking Free from the Cloud: Distributed, collaborative video encoding

James Le

Multimodal Video Understanding: How it's evolved and where it's headed

May 26, 2023 1:30-2:30 PST
Watch now