Build Enterprise Generative AI Apps Using Llama 3 at 1,000 Tokens/s on the SambaNova AI Platform
AI Engineer via YouTube
Gain a Splash of New Skills - Coursera+ Annual Nearly 45% Off
Introduction to Programming with Python
Overview
Coursera Spring Sale
40% Off Coursera Plus Annual!
Grab it
Learn to build enterprise-grade generative AI applications using Llama 3 at unprecedented speeds of 1,000 tokens per second on the SambaNova AI platform in this 55-minute intermediate-level workshop. Discover SambaNova's full-stack generative AI platform powered by the SN40L AI chip and explore Samba-1, a trillion parameter composition of experts model designed for enterprise settings. Build and deploy a complete question-answering application with retrieval augmented generation (RAG) for enterprise search using a comprehensive technology stack including LangChain framework, Unstructured for text preprocessing, E5-large-v2 embedding, ChromaDB vector store, and Llama-3-8B-Instruct. Gain hands-on experience through step-by-step Jupyter notebooks and Streamlit applications while working with provided SambaNova API keys for both CoE and Llama-3 endpoints. Master the integration of cutting-edge AI hardware with practical software frameworks to create high-performance enterprise AI solutions suitable for real-world deployment scenarios.
Syllabus
Build enterprise generative AI apps using Llama 3 at 1,000 tokens/s on the SambaNova AI platform
Taught by
AI Engineer