logo
episode-header-image
Jun 2024
45m 41s

Long Context Language Models and their B...

Sam Charrington
About this episode

Today, we're joined by Eric Nguyen, PhD student at Stanford University. In our conversation, we explore his research on long context foundation models and their application to biology particularly Hyena, and its evolution into Hyena DNA and Evo models. We discuss Hyena, a convolutional-based language model developed to tackle the challenges posed by long context lengths in language modeling. We dig into the limitations of transformers in dealing with longer sequences, the motivation for using convolutional models over transformers, its model training and architecture, the role of FFT in computational optimizations, and model explainability in long-sequence convolutions. We also talked about Hyena DNA, a genomic foundation model pre-trained on 1 million tokens, designed to capture long-range dependencies in DNA sequences. Finally, Eric introduces Evo, a 7 billion parameter hybrid model integrating attention layers with Hyena DNA's convolutional framework. We cover generating and designing DNA with language models, hallucinations in DNA models, evaluation benchmarks, the trade-offs between state-of-the-art models, zero-shot versus a few-shot performance, and the exciting potential in areas like CRISPR-Cas gene editing.


The complete show notes for this episode can be found at https://twimlai.com/go/690.

Up next
Aug 19
Genie 3: A New Frontier for World Models with Jack Parker-Holder and Shlomi Fruchter - #743
Today, we're joined by Jack Parker-Holder and Shlomi Fruchter, researchers at Google DeepMind, to discuss the recent release of Genie 3, a model capable of generating “playable” virtual worlds. We dig into the evolution of the Genie project and review the current model’s scaled-u ... Show More
1h 1m
Aug 12
Closing the Loop Between AI Training and Inference with Lin Qiao - #742
In this episode, we're joined by Lin Qiao, CEO and co-founder of Fireworks AI. Drawing on key lessons from her time building PyTorch, Lin shares her perspective on the modern generative AI development lifecycle. She explains why aligning training and inference systems is essentia ... Show More
1h 1m
Jul 29
Context Engineering for Productive AI Agents with Filip Kozera - #741
In this episode, Filip Kozera, founder and CEO of Wordware, explains his approach to building agentic workflows where natural language serves as the new programming interface. Filip breaks down the architecture of these "background agents," explaining how they use a reflection lo ... Show More
46m 1s
Recommended Episodes
May 2024
Using Single-Cell Multiomics to Characterize Human Developmental Hematopoiesis (Ana Cvejic)
In this episode of the Epigenetics Podcast, we talked with Ana Cvejic from the Biotech Research & Innovation Centre at the University of Copenhagen about her work on using sc-multiomics to characterise human developmental hematopoiesis. The conversation starts by delving into Ana ... Show More
37m 12s
Jul 2021
A life-changing database
Proteomes, the sequences of protein within the DNA of every living thing, are notoriously difficult to model. The usual chemical methods can take months, but a new computational model using the ability of artificial intelligence to learn the complex sequences is able to predict s ... Show More
32m 51s
Jul 2023
MosaicML's Naveen Rao on Making Custom LLMs More Accessible - Ep. 199
Startup MosaicML is on a mission to help the AI community enhance prediction accuracy, decrease costs, and save time by providing tools for easy training and deployment of large AI models. In this episode of NVIDIA's AI Podcast, host Noah Kravitz speaks with MosaicML CEO and co-f ... Show More
31m 26s
May 2022
How Garvan Institute of Medical Research is Driving Digital Frontiers in Genomics Analytics
In this episode, Associate Professor Sarah Kummerfeld, Head of Data Science for the Garvan Institute of Medical Research, joins Jay and Theo to discuss the vital role cloud computing plays in genomic sequencing. Genomic sequencing has been contributing to medical research to impr ... Show More
20 m
Sep 2023
Anima Anandkumar on Using Generative AI to Tackle Global Challenges - Ep. 203
Generative AI-based models can not only learn and understand natural languages — they can learn the very language of nature itself, presenting new possibilities for scientific research. Anima Anandkumar, Bren Professor at Caltech and senior director of AI research at NVIDIA, was ... Show More
40m 8s
Jul 2023
#130 Mathew Lodge: The Future of Large Language Models in AI
Welcome to episode #130 of Eye on AI with Mathew Lodge. In this episode, we explore the world of reinforcement learning and code generation. Mathew Lodge, the CEO of Diffblue, shares insights into how reinforcement learning fuels generative AI. As we explore the intricacies of re ... Show More
49m 44s
Jan 2024
BET Proteins and Their Role in Chromosome Folding and Compartmentalization (Kyle Eagen)
In this episode of the Epigenetics Podcast, we talked with Kyle Eagen from Baylor College of Medicine about his work on BET Proteins and their role in chromosome folding and compartmentalization. In the early days of his research career Dr. Eagen made use of genomics and microsco ... Show More
30m 53s
Jun 2024
Animal Computer Interaction
You’ve heard of Human Computer Interaction (HCI), now get ready for Animal Computer Interaction (ACI). Ilyena has made a career developing computer interfaces for non-human animals. She has worked with dogs, parrots, primates, and even giraffes. This is challenging because animal ... Show More
42m 49s
Aug 2022
Dr. Erich Jarvis: The Neuroscience of Speech, Language & Music
My guest this episode is Dr. Erich Jarvis, PhD—Professor and the Head of the Laboratory of Neurogenetics of Language at Rockefeller University and Investigator with the Howard Hughes Medical Institute (HHMI). Dr. Jarvis’ research spans the molecular and genetic mechanisms of voca ... Show More
1h 51m