Episode 183 - Gemini 1.5: One Million Tokens, Endless Possibilities?

Published: Feb. 23, 2024, 4 p.m.

Google's Gemini 1.5 is here, boasting a mind-blowing 1 million token context window! \U0001f92f Join Allen and Linda as they dive deep into this experimental AI, exploring its capabilities, limitations, and potential use cases. \U0001f914

\n


\n

They share their experiences testing Gemini 1.5 with original content, including Two Voice Devs transcripts and synthetic videos, and discuss the challenges of finding data that hasn't already been used to train the AI. \U0001f9d0

\n


\n

Get ready for a lively discussion on hallucinations, the future of content creation, and the ethical questions surrounding these powerful language models. \U0001f916

\n


\n

More info:

\n

* https://blog.google/technology/ai/google-gemini-next-generation-model-february-2024/

\n

* https://developers.googleblog.com/2024/02/gemini-15-available-for-private-preview-in-google-ai-studio.html

\n

* https://openai.com/sora

\n


\n

Timestamps:

\n

00:00:00 Introduction

\n

00:01:05 Notable features of Gemini 1.5

\n

00:02:57 What is a token?

\n

00:06:39 Linda's test with Danish citizenship PDF

\n

00:09:33 Allen's test with Les Miserables and needle in a haystack

\n

00:12:27 Testing with Data Portability API data

\n

00:14:28 Linda's test with YouTube search history and Netflix recommendations

\n

00:17:44 Allen's test with Two Voice Devs transcripts

\n

00:21:32 Issues with counting and hallucinations

\n

00:24:21 Testing with OpenAI's Sora AI synthetic videos

\n

00:30:05 Ethical questions and the future of content creation

\n

00:31:50 Potential use cases for large context windows

\n

00:36:34 API limitations and challenges

\n

00:37:39 Performance and cost considerations

\n

00:41:34 Comparison with retrieval augmented generation and vector databases

\n

00:44:21 Generating summaries and markers from this transcript

\n


\n

Leave your thoughts and questions in the comments below!