Further Notes on Our Recent Research on AI Delegation and Long-Horizon Reliability

Our recent paper, “LLMs Corrupt Your Documents When You Delegate”, has generated discussion about the reliability of AI systems in delegated workflows. We appreciate the interest in this work and want to clarify several important points about what the paper does—and does not—claim. The research aims to develop robust evaluation methods for long-horizon delegated and collaborative tasks. More broadly, this work reflects an ongoing effort to better understand the gap between strong benchmark performance and certain real-world tasks. Using a […]

Read more

Quiz: Python’s Array: Working With Numeric Data Efficiently

Interactive Quiz ⋅ 12 QuestionsBy Joseph Peart Share In this quiz, you’ll test your understanding of Python’s Array: Working With Numeric Data Efficiently. By working through this quiz, you’ll revisit the differences between Python’s array module and the built-in list, the meaning of type codes, how to create and manipulate arrays as mutable sequences, and the performance trade-offs of using a low-level numeric container. The quiz contains 12 questions and there is no time limit. You’ll get 1 point for […]

Read more

Quiz: Python Metaclasses

Interactive Quiz ⋅ 8 QuestionsBy Joseph Peart Share In this quiz, you’ll test your understanding of Python Metaclasses. Metaclasses sit behind every class you write in Python, and they’re one of the language’s deeper object-oriented concepts. By working through this quiz, you’ll revisit how classes are themselves objects, how type creates them, and how a custom metaclass lets you customize class creation. You’ll also reflect on when a custom metaclass is actually the right tool and when a simpler technique […]

Read more

Quiz: Cursor vs Windsurf: Which AI Code Editor Is Best for Python?

Interactive Quiz ⋅ 10 QuestionsBy Joseph Peart Share In this quiz, you’ll test your understanding of Cursor vs Windsurf: Which AI Code Editor Is Best for Python? By working through these questions, you’ll revisit how the two editors differ across code completion, agentic multi-file editing, and debugging. You’ll also reconnect with the audit points worth applying whenever an AI agent writes Python on your behalf. The quiz contains 10 questions and there is no time limit. You’ll get 1 point […]

Read more

Quiz: Cursor vs Windsurf: Which AI Code Editor Is Best for Python?

Interactive Quiz ⋅ 10 QuestionsBy Joseph Peart Share In this quiz, you’ll test your understanding of Cursor vs Windsurf: Which AI Code Editor Is Best for Python? By working through these questions, you’ll revisit how the two editors differ across code completion, agentic multi-file editing, and debugging. You’ll also reconnect with the audit points worth applying whenever an AI agent writes Python on your behalf. The quiz contains 10 questions and there is no time limit. You’ll get 1 point […]

Read more

How to Use OpenCode for AI-Assisted Python Coding

OpenCode is an open-source AI coding agent that runs in your terminal and lets you analyze and refactor a Python project through conversational commands. In this guide, you’ll install it on your system, set it up with a free Google Gemini API key, and learn the basics of how to use it in your daily programming work. Here’s what OpenCode’s main interface looks like: OpenCode’s Initial Screen OpenCode works as a conversational assistant you explicitly direct. Ask it to   […]

Read more

Quiz: The LEGB Rule & Understanding Python Scope

Interactive Quiz ⋅ 8 QuestionsBy Joseph Peart Share In this quiz, you’ll test your understanding of The LEGB Rule & Understanding Python Scope. By working through this quiz, you’ll revisit how Python resolves names using the LEGB rule, what the local, enclosing, global, and built-in scopes look like in practice, and how the global and nonlocal statements let you reach across scope boundaries. The quiz contains 8 questions and there is no time limit. You’ll get 1 point for each […]

Read more

Quiz: Building Type-Safe LLM Agents With Pydantic AI

Interactive Quiz ⋅ 9 QuestionsBy Joseph Peart Share In this quiz, you’ll test your understanding of Building Type-Safe LLM Agents With Pydantic AI. By working through this quiz, you’ll revisit how Pydantic AI returns structured outputs from LLMs, how validation retries improve reliability, how tools and function calling work, how dependency injection flows through RunContext, and what trade-offs to expect when running agents in production. The quiz contains 9 questions and there is no time limit. You’ll get 1 point […]

Read more
1 2 3 1,036