Dieter Flick

Building Real-Time AI/ML Agents with Distributed Data using Apache Cassandra and Astra DB

Are your LLMs stuck on outdated data? Learn how the RAG pattern and a vector database can build smarter, context-aware AI agents.

Building Real-Time AI/ML Agents with Distributed Data using Apache Cassandra and Astra DB
#1about 3 minutes

Introducing the DataStax real-time data cloud

The platform combines Apache Cassandra, Apache Pulsar, and Kaskada to provide a flexible database, streaming, and machine learning solution for developers.

#2about 3 minutes

Interacting with Astra DB using GraphQL and REST APIs

A live demonstration shows how to create a schema, ingest data, and query tables in Astra DB using both GraphQL and REST API endpoints.

#3about 1 minute

Understanding real-time AI and its applications

Real-time AI leverages the most recent data to power predictive analytics and automated actions, as seen in use cases from Uber and Netflix.

#4about 2 minutes

What is Retrieval Augmented Generation (RAG)?

RAG is a pattern that allows large language models to access and use your proprietary, up-to-date data to provide contextually relevant responses.

#5about 3 minutes

Key steps for building a generative AI agent

The process involves defining the agent's purpose, choosing an LLM, selecting context data, picking an embedding model, and performing prompt engineering.

#6about 3 minutes

Exploring the architecture of a RAG system

A RAG system uses a vector database to perform a similarity search on data embeddings, finding relevant context to enrich the prompt sent to the LLM.

#7about 3 minutes

Generating vector embeddings from text content

A Jupyter Notebook demonstrates splitting source text into chunks and using an embedding model to create vector representations for storage and search.

#8about 4 minutes

The end-to-end data flow of a RAG query

A user's question is converted into an embedding, used for a similarity search in the vector store, and the results are combined with other context to build a final prompt.

#9about 3 minutes

Executing a RAG prompt to get an LLM response

The demo shows how the context-enriched prompt is sent to an LLM to generate a relevant answer, including how to add memory for conversational history.

#10about 3 minutes

Getting started with the Astra DB vector database

Resources are provided for getting started with Astra DB, including quick starts, a free tier for developers, and information on multi-cloud region support.

Related jobs
Jobs that call for the skills explored in this talk.

Featured Partners

Related Articles

View all articles
DC
Daniel Cranney
Dev Digest 213: Petrol Prices, Agentic Workflows, AI Skills and CODE100!
Inside last week’s Dev Digest 213 . 🤫 Don’t tell your LLM that it is an expert 👻 AI generated code is invisible 🔄 Learn about agentic workflows 🛡️ Linux Foundation sponsors fight against AI slop 🦠 1M users infected by Chrome extension 🫃 The why of J...
 Dev Digest 213: Petrol Prices, Agentic Workflows, AI Skills and CODE100!

From learning to earning

Jobs that call for the skills explored in this talk.

AI Security Engineer

Databricks
Amsterdam, Netherlands

Intermediate
C++
Python
PyTorch
TensorFlow
Machine Learning