Skip to content

AI Engineering

๐Ÿš€ The Future of Engineering & AI โ€” Insights from CEOs

Michael Truell - CEO of Cursor

"We're not going straight to a world where AI does everything and engineers disappear. Instead, engineers are shifting rolesโ€”from implementers to orchestrators."

Everyone Become Engineering Managers

"I think something people don't talk enough about when discussing AI agents and AI engineers doing all this stuff for youโ€ฆ is basically we're all becoming engineering managers."

Why Reinforcement Learning (RL) is hot again?

Just finished listening to an incredible podcast featuring an interview with Wu Yi โ€” a Tsinghua alum and former OpenAI researcher โ€” and his take on Reinforcement Learning (RL) was one of the clearest Iโ€™ve seen!

๐Ÿ” 1. What is RL really about?

Wu Yi explains that RL is very different from traditional supervised learning (like image classification). In supervised learning, we train models using a fixed set of labeled data โ€” one-shot answers.

RL, on the other hand, is more like playing a game: you need to make a sequence of decisions (serve, move, react), and there's no single โ€œcorrectโ€ path. The quality of your decisions is judged by the final outcome (win or lose). Itโ€™s about multi-step decision-making โ€” much closer to how the real world works.

๐Ÿค– 2. Why is RL hot again? Whatโ€™s its connection to LLMs?

Prompt Engineering

Content


Key Takeaways:
Prompt engineering is an ๐—ถ๐˜๐—ฒ๐—ฟ๐—ฎ๐˜๐—ถ๐˜ƒ๐—ฒ ๐—ฝ๐—ฟ๐—ผ๐—ฐ๐—ฒ๐˜€๐˜€ involving continuous ๐˜๐—ฒ๐˜€๐˜๐—ถ๐—ป๐—ด, ๐—บ๐—ผ๐—ฑ๐—ถ๐—ณ๐—ถ๐—ฐ๐—ฎ๐˜๐—ถ๐—ผ๐—ป, ๐—ฎ๐—ป๐—ฑ ๐—ผ๐—ฝ๐˜๐—ถ๐—บ๐—ถ๐˜‡๐—ฎ๐˜๐—ถ๐—ผ๐—ป. As models continue to evolve, prompts must adapt accordingly to maintain effectiveness.


What is prompt engineering?

Prompt engineering is about "communicating" with LLM in a way that maximizes the model's understanding and performance on a given task. At its core, prompt engineering involves designing, refining, and optimizing the text inputs (prompts) given to models to elicit accurate, relevant, and useful responses.

Playwright vs Puppeteer vs Selenium

In todayโ€™s fast-paced development landscape, automated browser testing and web scraping are more vital than ever. Three of the most prominent tools in this space are Selenium, Puppeteer, and Playwright. Each has its unique strengths and target use cases. In this post, weโ€™ll explore how they stack up in terms of browser support, language options, performance, ease of use, and more.

However, for the most powerful and reliable option, Playwright is the best of the three.

Top best practices for building production-ready AI apps

  1. Build evals
  2. Define test cases to ensure you're actively improving your app & not causing any regressions.

  3. Break down one LLM call into multiple

  4. AI systems do a lot better when you have many LLM calls chained together. i.e, instead of sending an LLM call to a model to generate code, send it to a "architect" model to generate a plan, then a "coding" model to generate code, then a "reviewer" model to verify.

  5. Start simple (with 1 LLM call)

  6. Then iterate with prompt engineering (few shot examples, chain of thought, descriptive prompts) before building a more complex system with chained LLM calls.

Cursor AI Done Right: Lessons from Building Multiple MVPs

Cursor is really dumb if not given enough context about your project. Here what you can do to improve your Cursor workflow

1. Brainstorm first, code second

Claude/o1 are your best friends here. You should create a whole document containing every single detail of your project.

  • core features
  • goals & objectives
  • tech stack & packages
  • project folder structure
  • database design
  • landing page components
  • color palette
  • copywriting

All this should be put into an instruction.md (name it however you want) so Cursor can index at any time.

What's Model Context Protocol (MCP)

Introduction

Model Context Protocol (MCP) is an open protocol that standardizes how applications provide context to LLMs. It's a new standard for connecting AI assistants to the systems where data lives, including content repositories, business tools, and development environments. Its aim is to help frontier models produce better, more relevant responses.

Why MCP?

As AI assistants gain mainstream adoption, the industry has invested heavily in model capabilities, achieving rapid advances in reasoning and quality. Yet even the most sophisticated models are constrained by their isolation from dataโ€”trapped behind information silos and legacy systems. Every new data source requires its own custom implementation, making truly connected systems difficult to scale.

MCP addresses this challenge. It provides a universal, open standard for connecting AI systems with data sources, replacing fragmented integrations with a single protocol. The result is a simpler, more reliable way to give AI systems access to the data they need. โ€‹

What is RAG?

Introduction๐Ÿš€

RAG is a popular method that improves accuracy and relevance by finding the right information from reliable sources and transforming it into useful answers.

Large Language Models are trained on a fixed dataset, which limits their ability to handle private or recent information. They can sometimes "hallucinate", providing incorrect yet believable answers. Fine-tuning can help but it is expensive and not ideal for retraining again and again on new data. The Retrieval-Augmented Generation (RAG) framework addresses this issue by using external documents to improve the LLM's responses through in-context learning. RAG ensures that the information provided by the LLM is not only contextually relevant but also accurate and up-to-date.

final diagram

There are four main components in RAG: