{ "cells": [ { "cell_type": "markdown", "id": "4a6ab9a2-28a2-445d-8512-a0dc8d1b54e9", "metadata": {}, "source": [ "# Code Generator\n", "\n", "The requirement: use a Frontier model to generate high performance C++ code from Python code\n" ] }, { "cell_type": "markdown", "id": "d5ccb926-7b49-44a4-99ab-8ef20b5778c0", "metadata": {}, "source": [ "
\n",
" \n",
" | \n",
" \n",
" Reminder: fetch latest code\n", " I'm continually improving these labs, adding more examples and exercises.\n", " At the start of each week, it's worth checking you have the latest code.\n", " First do a git pull and merge your changes as needed. Any problems? Try asking ChatGPT to clarify how to merge - or contact me! \n", " After you've pulled the code, from the llm_engineering directory, in a Cursor Terminal, run: \n", " uv sync\n", " \n", " | \n",
"
\n",
" \n",
" | \n",
" \n",
" Important Note\n", " \n", " In this lab, I use high end models GPT 5, Claude 4.5 Sonnet, Gemini 2.5 Pro, Grok 4, which are the slightly higher priced models. The costs are still low, but if you'd prefer to keep costs ultra low, please pick lower cost models like gpt-5-nano.\n", " \n", " | \n",
"