Building Real-Time AI/ML Agents with Distributed Data using Apache Cassandra and Astra DB
Are your LLMs stuck on outdated data? Learn how the RAG pattern and a vector database can build smarter, context-aware AI agents.
#1about 3 minutes
Introducing the DataStax real-time data cloud
The platform combines Apache Cassandra, Apache Pulsar, and Kaskada to provide a flexible database, streaming, and machine learning solution for developers.
#2about 3 minutes
Interacting with Astra DB using GraphQL and REST APIs
A live demonstration shows how to create a schema, ingest data, and query tables in Astra DB using both GraphQL and REST API endpoints.
#3about 1 minute
Understanding real-time AI and its applications
Real-time AI leverages the most recent data to power predictive analytics and automated actions, as seen in use cases from Uber and Netflix.
#4about 2 minutes
What is Retrieval Augmented Generation (RAG)?
RAG is a pattern that allows large language models to access and use your proprietary, up-to-date data to provide contextually relevant responses.
#5about 3 minutes
Key steps for building a generative AI agent
The process involves defining the agent's purpose, choosing an LLM, selecting context data, picking an embedding model, and performing prompt engineering.
#6about 3 minutes
Exploring the architecture of a RAG system
A RAG system uses a vector database to perform a similarity search on data embeddings, finding relevant context to enrich the prompt sent to the LLM.
#7about 3 minutes
Generating vector embeddings from text content
A Jupyter Notebook demonstrates splitting source text into chunks and using an embedding model to create vector representations for storage and search.
#8about 4 minutes
The end-to-end data flow of a RAG query
A user's question is converted into an embedding, used for a similarity search in the vector store, and the results are combined with other context to build a final prompt.
#9about 3 minutes
Executing a RAG prompt to get an LLM response
The demo shows how the context-enriched prompt is sent to an LLM to generate a relevant answer, including how to add memory for conversational history.
#10about 3 minutes
Getting started with the Astra DB vector database
Resources are provided for getting started with Astra DB, including quick starts, a free tier for developers, and information on multi-cloud region support.
Related jobs
Jobs that call for the skills explored in this talk.
Panel Discussion: Responsible AI in Practice - Real-World Examples and ChallengesIntroductionIn the ever-evolving landscape of artificial intelligence, the concept of "responsible AI" has emerged as a cornerstone for ethical and practical AI implementation. During the WWC24 Panel discussion, three eminent experts—Mina, Bjorn Brin...
Dev Digest 215: Agent Memory, JS2026, Googlebot Analysis & Canvas❤️HTMLInside last week’s Dev Digest 215 .
🗿 Make AI talk like a caveman
🧠 A guide to context engineering for LLMs
🤖 Simon Willison on agentic engineering
🔐 Axios supply chain attack post mortem
🛡️ Designing AI agents to resist prompt injection
🎨 HTML in c...
From learning to earning
Jobs that call for the skills explored in this talk.