Evaluating LLM Performance: DeepSeek, Phi-3.5, and LLaMA 3 Using Chain of Thought Reasoning
The Machine Learning Engineer via YouTube
Google AI Professional Certificate - Learn AI Skills That Get You Hired
MIT Sloan AI Adoption: Build a Playbook That Drives Real Business ROI
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore a detailed comparison of logical and mathematical reasoning capabilities across three language models - DeepSeek R1 1.5B, Microsoft Phi3.5 3.8B, and LLama 3.2 3B - in this 26-minute technical evaluation video. Learn how to implement Chain of Thought (CoT) prompting techniques using Ollama and Langchain frameworks, with all models quantized to 4Int for performance analysis. Access the complete evaluation process through the provided Jupyter notebook to understand the comparative strengths and limitations of these models in handling complex reasoning tasks.
Syllabus
LLM,s: CoT Evaluation DeepSeek R1 1.B, Phi3.5 3.8B and LLama 3.2 3B #datascience #machinelearning
Taught by
The Machine Learning Engineer