This page contains notes, further readings, and general references for the associated lecture.
The Lecture Slides for the talk are also available (use your arrow keys to advance the slides) or download a pdf of the slides.
If you’re looking for notes to the other lectures, head here.
Reliability and Hallucinations
We spent a good portion of the first hour discussing why models lie. The technical term is “hallucination,” and it stems from the probabilistic nature of the technology (as discussed in our previous lecture). A recent report by UX Tigers titled AI Hallucinations on the Decline gives a great breakdown of the current error rates (between 0.7% and 3%) for the major frontier models.
To understand why models sometimes lie to please you (Sycophancy), the researchers at Anthropic (the makers of Claude) released a fascinating paper titled Towards Understanding Sycophancy in Language Models. It is a bit academic, but the abstract and introduction are very readable and explain why an AI might agree with a user’s wrong opinion just to be “helpful.”
Here’s a great article from TIME focusing on the issues of AI sycophancy
GeorgeTown Law has compiled a large list of examples where AI sycophancy has proved harmful, and open questions that researchers still need to deal with to make AI models safer.
Synthetic Media Generation
We looked at the “Blind Test” experiment conducted by fantasy author Mark Lawrence. You can read his full blog post here where he details how audiences rated AI-generated stories higher than human ones. One suspected reason for this is that the AI is able to perfectly hit the “average” tropes the readers liked to see, whereas the human authors didn’t quite hit things correctly.
The “Grandparent Scam” has evolved significantly. McAfee released a 2025 Report on AI Fraud detailing how voice cloning technology has lowered the barrier for entry for scammers.
Remember, as discussed in our lecture, please establish a “Safe Word” with your family. This should be a word or phrase that you never share on social media. If you receive a distressed call from a loved one, ask for the word.
Education
While the “Calculator for Thought” analogy of AI is powerful, it maybe doesn’t capture the full scope. Still, it shows some promise. For a look at one positive potential of AI tutors, Khan Academy’s Khanmigo is a very reasonable and thought provoking implementation. They have excellent demo videos showing how the AI guides students rather than just giving answers.
On the flip side, MIT Media lab put out a pretty damning article how using AI tools early can literaly atrophy cognitive skills. Good news though, the group that started without AI tools, learned, then later adopted them did the overall best, which might be the best way forward.
Another interesting article (and quite recent) was put out by University World News examining the results of a nationwide survey of educators on the impact of AI on their students. This is less scientific but helps capture the feeling of most educators I have spoken with.
Biology
The 2024 Nobel Prize in Chemistry was awarded to Demis Hassabis and John Jumper of Google DeepMind for AlphaFold. You can read the official press release from the Nobel Committee which explains the significance of solving the “protein folding problem”, a 50-year grand challenge in biology that AI solved in roughly 18 months.
The original Nature paper detailing the release of AlphaFold
A full review article of the protein folding story appeared in Quanta and is just a fantastic read. Highly recommended.
Employment
- We referenced the Goldman Sachs report “How Will AI Affect the Global Workforce?” during our lecture. It is heavy on economics, but the charts are very clear.
Robotics
- You can see the Figure 03 robot launch video where it demonstrates how it can fold clothes, put away dishes, and clean up a room. It’s definitely not perfect (see this video by Time for a more realistic take), but the speed of progress is undeniable.
Consciousness and Agency
- We ended on a philosophical note. The debate on whether these models are “conscious” or just “mimicking consciousness” is ongoing. The research paper we discussed, written by researches at Anthropic, give us some surprising early signs that these models can indeed be introspective about their own internal states, which is a common part of most definitions of “consciousness”.