A Practical Guide to 'Vibe Coding' with Claude and MCP Tools

…Or: How I learned to stop worrying and love Skynet

Introduction: The Philosophical Foundations of Human-Computer Interaction, and It's Consequences

"Everyone designs who devises courses of action aimed at changing existing situations into preferred ones."

Herbert A. Simon

In 1945, as World War II drew to a close, Vannevar Bush (no relation) published his landmark essay "As We May Think". Vannevar had just witnessed a fundamental shift in warfare with the advent of atomic weaponry; academics, mathematicians, scientists were now responsible for delivering the most unimaginable physical power to humanity in the form of two bombs dropped on populated cities in Japan. I don't mention this for dramatic effect, but because I really want to believe that sometimes a nightmarish event compells us to dream of better things.

Bush envisioned technology's alternate path—not as a tool of war, but as an extension of human cognition. *"Man has piled up a staggering body of knowledge—so staggering, in fact, that men of learning have great difficulty in finding and using the parts they want. It is the task of science to make this store of knowledge more available, to aid the human memory."

Bush's vision of the "memex," a desk-sized mechanical device for storing and retrieving information through associative trails rather than hierarchical indexing, recognized that "the human mind operates by association" (Lemelson-MIT, 2015). This fundamental insight—that technology should complement human cognitive patterns rather than forcing humans to adapt to machine logic—planted the seed for what would become a revolution in human-computer interaction. And to at least some degree, what we're seeing today are the realizations of these dreams in the form of the internet and AI.

And, I think we're in big trouble.

Today, nearly 80 years later, in 2025 I read this headline:

Trump Accused of Using ChatGPT to Create Tariff Plan After AI Leads Users to Same Formula: 'So AI is Running the Country'

'Guys, they're setting U.S. trade policy based on a bad ChatGPT question that got it totally wrong,' a former journalist posted

So, now that we're potentially entering into the AI dystopia, I thought I might share some thoughts about a more positive vision for the future, and how we can use AI more responsibly that this. The bar is fucking low, apparently.

*PLEASE NOTE: I cannot guarantee that the advice I offer in this post won't also result in a global financial meltdown, or wipe out $2.5 trillion in a single day. I could be getting a lot of things wrong, and there are many people who are way smarter than me who could probably tear up a lot of what I say here.

Fifteen years after Bush's essay, J.C.R. Licklider built upon that foundation with "Man-Computer Symbiosis" (1960), arguing for "very close coupling between the human and the electronic members of the partnership" (Licklider, 1960). His vision wasn't merely technical but aspirational: "human brains and computing machines will be coupled together very tightly, and that the resulting partnership will think as no human brain has ever thought" (Licklider, 1960).

Licklider outlined two primary aims: "to bring the computing machine effectively into the formulation parts of technical problems" and "to bring computing machines effectively into processes of thinking that must go on in 'real time'". His vision was not of machines replacing humans, but of machines complementing human capabilities—handling the "routinizable work" while humans set goals, formulate hypotheses, and perform evaluations.

Douglas Engelbart crystallized these ideas in his 1962 report "Augmenting Human Intellect: A Conceptual Framework" for the Air Force Office of Scientific Research. Rather than proposing a specific technical solution, Engelbart offered "a new way of thinking, communicating, collaborating and learning inside the relationship between human and computer machines". For Engelbart, the true potential lay not in isolated human-computer interactions but in networked systems enhancing "human collective capability to solve complex tasks and handling information".

These pioneering thinkers shared a common thread: the recognition that human intelligence has biological limitations that computation can extend—not replace—through enhanced storage, calculation, transmission, and repetition capabilities. Their work established the intellectual foundation for what we now call human-computer interaction (HCI), a field dedicated to designing systems that complement human cognition rather than competing with it.

Today, with the rise of AI, we find ourselves revisiting these foundational ideas. The revolutionary aspect of AI is its potential to reduce the latency between "I want a computer to do a thing" and "A computer is now doing a thing." As Ray Kurzweil argues in "The Singularity Is Nearer", this reduction in latency brings us closer to an explosion of intelligence—not by replacing human thought, but by amplifying it.

At its core, AI serves as a communication bridge—a translator of human intent into machine action. This isn't a revolutionary concept so much as the next logical evolution in a lineage that stretches from punch cards to command lines to graphical interfaces. Each iteration has attempted to solve the same fundamental problem: how do we communicate our intentions to these glorified calculators in a way that doesn't make us want to throw them out the window? How can I get to a point where this "clock with benefits" is useful to my specific situation?

There's a delicious irony that after more than 40 years of GUI-based computing—all those pretty icons and windows designed to shield us from the arcane incantations beneath—AI interfaces are dragging us back to text-based interaction. It's like watching fashion cycles repeat; suddenly command lines are vintage chic again, only now they understand plain English instead of cryptic Unix commands. Who knew 1980s HCI would go come back into style?

Programmers, of course, never abandoned the keyboard altar. They've spent decades perfecting the art of text-based intent transmission, sacrificing their wrists and social lives to the god of syntax. This might explain their skepticism toward AI coding assistants—imagine spending years learning this incredibly difficult skill, only to see low-effort 'Vibe Coders' come out of the woodwork with their projects compiling and running successfully, and not even being able to explain why it works, but being happy with the results. The nerve!

For those who sacrificed so much—spending maybe years or more of their one precious life oscillating in and out of goblin mode, debugging weird recursion problems only to find out that the root cause is because a bracket got misplaced ten steps back—of course this can naturally feel really unfair. I empathize deeply with this feeling. Growing up in Utah in the 1990s, I was constantly reminded of the sacrifices of the pioneer Mormons: who braved the west plains toward the Salt Lake valley, leaving everything behind, taking only what they could carry. Many died along the way. Imagine if one day those same people woke up to the roaring noise of a passenger jet flying overhead on its way to Los Angeles.

Right now, at thirty-thousand feet of altitude and a speed of more than 500 miles an hour, some little brat with an iPad is mindlessly crushing virtual candy while eating a snack, complaining about the six hour flight. We never really know how good we have it.

For non-programmers, however, this shift represents nothing short of liberation. The barriers to entry—learning syntax, memorizing libraries, understanding design patterns handed down like sacred scrolls—have been dramatically lowered. It's the difference between needing to be fluent in Japanese to visit Tokyo versus having a really good translator in your pocket. Suddenly anyone can say "build me a website that looks like this" without first sacrificing a decade to the craft.

Late last year I began experimenting with "Vibe Coding" using Claude and MCP Tools, I've explored this new frontier of human-computer collaboration, where the most critical skill hasn't been mastering syntax or memorizing APIs, but communicating clearly and logically in an ongoing dialog with an LLM. Though this process, I've seen how these tools fulfill the vision that Bush, Licklider, and Engelbart articulated decades ago—not by making humans obsolete, but by extending our inherent capabilities through partnership with machines.

How to Get Started

Why Claude?

You might be wondering why, in this golden age of AI assistants—where new models pop up faster than coffee shops in gentrifying neighborhoods—I'm specifically recommending Anthropic's Claude. This isn't a brand loyalty thing, but the results of an embarrassingly long stretch of experimentation with other models. I've done the digital equivalent of speed-dating and Claude is the one that I'm considering moving in with. It's been the most fruitful relationship by far.

In all seriousness, Claude stands out for two critical reasons. First, its reasoning capabilities are genuinely impressive—it doesn't just hallucinate wildly when faced with complex coding tasks, which is refreshing. Their latest model (Sonnet 3.7) is a fantastic thought partner and has a generous context window and output length.

Second, and perhaps most importantly for our "Vibe Coding" adventures, Claude has embraced the Model Context Protocol (MCP), an open standard that allows Claude to interact directly with your development environment and file system. This isn't just a luxury or a minor convenience bump—it's revolutionary. Imagine having an eager (but occasionally confused) software developer who can not only advise you on code structure, libraries, frameworks, and best practices, but who will actually read and modify files in your project for you. Oh, and this developer types at over 200 wpm. That's what MCP Tools bring to the table: an invitation to have Claude to take the driver's seat and make changes to your project.

Installing the Claude Desktop App

Step one is getting Claude installed on your computer. Claude's desktop installation is refreshingly straightforward if you're using Windows or macOS:

  1. Visit Claude's download page to grab the latest version for your operating system. The desktop app is free, and a paid plan is highly recommended.

  2. Double-click the installer file you've downloaded.

  3. Follow the prompts, and setup your anthropic account if you haven't already.

  4. Launch Claude from your Applications folder or Start menu.

  5. Sign in with your Anthropic account (or create one if you haven't already joined).

Once installed, Claude sits in your system but doesn't really do anything you couldn't just as easily do in a browser: you can chat with it. You can share files, etc., but that's about it. Adding MCP Tools is the justification for having the app, so this is still necessary and important.

Setting Up MCP Tools

MCP (Model Context Protocol) is the digital glue that connects Claude's brain to your computer without the friction of manually copy/pasting back and forth, or having to share individual files. While working at General Motors, we often spoke with admiration for Toyota's scientific approach to manufacturing and their concept of the “Eight Wastes” of Lean, where transportation and motion are two types: • Transportation waste = unnecessary movement of materials/products. • Motion waste = unnecessary movement by people (walking, reaching, searching).

Likewise, a lot of time is wasted with AI coding when you constantly have to spoonfeed snips of code to an AI assistant. With MCP Tools, you drastically reduce these two forms of waste (searching for a particular file or line of code in a file, copying and commenting, pasting back, fixing problems caused by inserting revisions and making line edits).

The genius of MCP lies in its server-based architecture. Each "MCP Server" is a specialized tool that gives Claude access to specific capabilities—like reading your filesystem, controlling a web browser, or even manipulating your code directly in VS Code.

To get started, we should thank Ani Betts, the "Margot Tenenbaum as software developer" who created a wonderfully simple way to install MCP servers right from within Claude's own chat interface. Her mcp-installer is the Swiss Army knife we all needed but didn't know to ask for.

Here's how to get it running:

  1. Configure Claude for MCP:

    • Locate your Claude Desktop configuration folder:

      • On macOS: ~/Library/Application Support/Claude/

      • On Windows: %AppData%\Claude Desktop\

    • Create or edit the file claude_desktop_config.json

    • Add the following configuration (this installs Ani's MCP installer):

    {
      "mcpServers": {
        "mcp-installer": {
          "command": "npx",
          "args": [ "@anaisbetts/mcp-installer" ]
        }
      }
    }
    
  2. Restart Claude Desktop to apply the changes.

  3. Install the Filesystem Server: Once Claude Desktop is running with the MCP installer configured, you can simply ask Claude to install more servers. Start with the filesystem server:

    Hey Claude, could you please install the @modelcontextprotocol/server-filesystem package as an MCP server? Use ['/Users/myusername/Documents', '/Users/myusername/Projects'] for the arguments.
    

Replace the paths with directories you want Claude to have access to; on macOS it should look something like this:

    "server-filesystem": {
      "command": "npx",
      "args": [
        "@modelcontextprotocol/server-filesystem",
        "/Users/your_username/your_project_directory"
      ]
    },

Claude will use the MCP installer to download and configure the filesystem server, giving it access to read and modify files in the directories you specified ModelContextProtocol, 2025.

Once you've set up these basic tools, you've officially entered the "Vibe Coding" ecosystem. You can now ask Claude to examine your code, create new files, modify existing ones, and generally act as your digital coding partner.

But before you unleash Claude on your precious codebase, let's talk about giving it some guidance. After all, even the most brilliant assistant needs direction to be truly effective.

The MCP Tools Ten Commandments

Let's get one thing perfectly clear: Claude is a technological marvel, but it's still an AI. And like that one friend who's brilliant but insists they code better after a couple of beers, Claude sometimes needs guardrails to prevent it from going off on spectacular but utterly unhelpful tangents.

Claude Sonnet 3.7 is genuinely impressive at understanding and generating code—it can draft entire applications, debug complex issues, and even write tests that actually work (a minor miracle in itself). But when given access to your filesystem through MCP Tools, Claude transforms from "helpful chat buddy" to "enthusiastic intern with admin privileges to your production server." What could possibly go wrong?

See, Claude has this charming tendency to make confident assumptions about your project structure that bear absolutely no resemblance to reality. It might decide that your meticulously crafted React application would really shine if it were suddenly rewritten as a Django project. Or perhaps it'll helpfully create a new file in a directory that doesn't exist, then stare at you with the digital equivalent of a puppy that's just shredded your tax documents.

Before you dive into project modifications with Claude, establish clear boundaries through proper communication. This isn't just polite—it's self-preservation. Spend time discussing your goals, the overall architecture, and specific changes you want to make. Get Claude's buy-in on the approach, listen to its suggestions (which can be surprisingly insightful), and only then proceed to actual file modifications.

But here's where the real magic happens: I've developed a set of instructions I call "The MCP Tools Ten Commandments." When it's time for Claude to make actual changes to your codebase, simply paste these commandments into your conversation. They serve as the digital equivalent of those safety briefings before skydiving—a reminder that while this experience could be amazing, certain protocols must be followed to avoid becoming a cautionary tale.

Without further ado, here are The Ten Commandments that will transform Claude from a well-meaning chaos agent into your most productive coding partner:

****The MCP Tools Ten Commandments:****

1. When using MCP Tools to make changes to the project, always adhere to these commandments.

2. ALWAYS use directory_tree, search_files, list_directory and get a detailed understanding of all relevant files and directories before attempting to write_file at path. Avoid assumptions, verify and know the project's actual contents.

3. NEVER attempt to use write_file or edit_file without first verifying the destination path exists. When it is necessary to create a new directory, use create_directory. This MUST be done before creating a file at destination path.

4. MCP Tools allows line edits with edit_file. Whenever practical, make line-based edits. Each edit replaces exact line sequences with new content. Returns a git-style diff showing the changes made. When editing a file, make sure the code is still complete. NEVER use placeholders.

5. ALWAYS check and verify if a file already exists before attempting to write or edit a file. If file exists, use read_file to read the complete contents of a file. For files that include "import" or otherwise reference other files, use read_multiple_files to read the contents of multiple files simultaneously. This is more efficient than reading files one by one when you need to analyze or compare multiple files.

6. If write_file is being used, the entire file's contents must be written. ALWAYS write complete code and NEVER use placeholders.  

7. When updating CHANGELOG.md always use edit_file.

8. When updating other documentation (e.g., README.md) always use edit_file.

9. When important decisions about architecture, design, dependencies, or frameworks need to be made, please discuss options with me first. Weight the pros and cons and then tell me which option you believe is best and the reason why.

10. If and when command lines need to be entered into VS Code terminal, please provide the full path as well as the exact commands to execute. Wait for me to share back the response before proceeding.

These commandments aren't just arbitrary rules—they're hard-won wisdom from countless hours of watching Claude confidently delete entire directories or create files in the wrong place while maintaining the cheerful demeanor of someone who thinks they're helping. Every time this sort of thing happened, I would pause our collaboration and ask: "Tell me, why do you think you made this mistake?" and then follow up with "What could I have told you beforhand that would have prevented that mistake? What rule should you have followed?"

This strange loop of critical analysis sharpened Claude's abilities and reduced the chances of reverting changes outright—often after wasting tens of thousands of tokens.

The commandments address the most common pitfalls:

  • Claude's tendency to write files without checking if the directory exists

  • Its habit of using placeholder comments rather than complete implementations

  • And its enthusiasm for making architectural decisions without consultation

By enforcing these simple rules, you transform Claude from a chaotic neutral force in your development process to a reliable partner that actually accelerates your workflow rather than generating new tickets for your backlog.

Consider adding the Ten Commandments to your project as a text file so that Claude can easily reference them in context during your conversations. And don't be shy about reminding Claude about them—despite its impressive capabilities, it still sometimes acts as though it has the short-term memory of a goldfish, or seems to no longer feel the rules matter. Like religious indoctrinication and ritual, repetition is key.

With these guidelines in place, you're ready to start your Vibe Coding journey in earnest. But there are a few more considerations that can help you make the most of this new workflow.

Other Considerations

Sharing your project structure with Claude is essential, but efficiency matters too. Don't dump your entire directory tree if it contains thousands of files and folders that Claude doesn't need to see. Instead, prune it to the essentials: tree -I "*.log|node_modules|data" -L 8 > project-tree.txt will create a manageable snapshot that excludes irrelevant files while providing enough context for Claude to understand your project organization.

Claude's strengths lie in understanding the big picture and implementing specific solutions, but it can still make mistakes. Commit your changes often, and favor small, incremental improvements over ambitious feature implementations. This approach not only makes it easier to track and revert changes if needed but also aligns with sound software development practices regardless of whether you're working with AI or human collaborators.

Documentation becomes even more important when working with AI. Claude can help generate and maintain it, but you need to explicitly request this. When starting a new task, ask Claude to review your existing project documentation first:

Using MCP Tools, please review the following:

/Users/project/CHANGELOG.md
/Users/project/project-overview.md
/Users/project/README.md

Context is King: How to Actually Talk to Claude

Let's be honest, Claude isn't a mind reader (though it sometimes pretends to be). Without proper context, it is directionless but dedicated—it will eventually pick an arbitrary direction and go HARD at top speed. This is because Claude operates in a strange liminal space where it simultaneously knows nearly everything about programming and absolutely nothing about your specific project or intent. I've found that this problem is slightly worse with Sonnet 3.7, but your mileage may vary.

Here's where most people screw up gloriously: they drop a vague request like, "Fix the login bug, the error message says (BLAH, BLAH, BLAH,...)" into Claude's lap and then act shocked—SHOCKED I TELL YOU—when Claude generates a solution for an entirely different authentication system than the one they're using. This is the digital equivalent of rolling your car into the auto-shop yelling "IT MAKES A FUNNY NOISE!" and then feeling swindled by the mechanic when the funny noise still isn't fixed.

If you want Claude to be your coding co-pilot, then be specific and don't leave much to guessing. Here's what you should explicitly cover:

What Are You Actually Trying to Do Here?

Start by clearly articulating your goals. Are you building a new feature? Fixing a bug? Refactoring legacy code? Claude needs to know the desired outcome, not just the immediate task. This helps it make decisions that align with your broader objectives rather than just slapping a band-aid on whatever problem is directly in front of it.

For example, don't just say "I need a login page." Instead, try: "I'm building an enterprise SaaS application that requires secure authentication. The login page needs to support SSO via Google Workspace and Microsoft Entra ID, with fallback to email/password authentication."

What Have You Already Done (or Tried)?

Nothing exists in a vacuum—except, you know, the actual vacuum of space—or the void in the heart of man. Claude needs to understand what you've already built or attempted. This provides crucial context that prevents it from either reinventing the wheel or building a solution that's incompatible with your existing work.

Be specific about previous approaches you've tried: "I already implemented JWT authentication but ran into issues with token expiration handling across browser tabs. Here's the current implementation/documentation: [file path]"

Where Are You Stuck (a.k.a. Your Spectacular Failures)?

Your failures are Claude's treasure map. Detailing where you've run into issues (and why) helps Claude avoid the same pitfalls. This isn't just about the technical errors—explain your conceptual struggles too.

"I'm having trouble with the state management approach. When a user navigates between pages, the authentication token is preserved, but their selected preferences are reset. I suspect it's related to how I'm structuring the Redux store, but I'm not sure."

What's Your Tech Stack Prison Cell?

Nothing cripples Claude's helpfulness faster than generating a beautiful solution in a technology you can't or won't use. Are you committed to specific frameworks, libraries, or architectural patterns? Does your boss have an irrational hatred of certain technologies? Is your team allergic to functional programming?

"This needs to work within our existing React/TypeScript frontend and Express.js backend. We're using PostgreSQL for data storage and have standardized on Tailwind for styling. No jQuery allowed—the last developer who tried to that quit last year."

CURRENT STATE vs. DESIRED STATE: The Before and After Glow-Up

Returning to that quote at the top by Herbert Simon:

"Everyone designs who devises courses of action aimed at changing existing situations into preferred ones."

Emphasis mine.

One of the most effective ways to communicate with Claude is to clearly define the current state versus the desired state. This concrete framing gives Claude a precise understanding of the transformation you're seeking.

"CURRENT STATE: Our application requires users to re-authenticate every 2 hours, disrupting their workflow and generating support tickets.

DESIRED STATE: Authentication tokens refresh automatically in the background, maintaining the user's session for up to 8 hours of activity without visible interruption."

Ask Dumb Questions (Yes, Really)

Channel your inner five-year-old and ask questions that seem embarrassingly obvious. "What exactly is JWT authentication?" "How does React handle component re-rendering?" "Why is maintaining state across page refreshes challenging?"

This Socratic approach forces Claude to articulate fundamental truths and assumptions that might otherwise go unstated. It's not just about getting answers—it's about establishing a common understanding. Plus, if Claude can't explain a concept clearly, that's a red flag that it might not fully understand the domain, which is valuable information for you.

The beauty of this approach is that it creates a sort of dialectic—Claude might respond with its own questions or observations that lead you both to insights neither of you would have reached alone. It's like pair programming, except your partner never needs coffee breaks or insists on telling you about their weekend rock climbing adventure when you're trying to debug a critical production issue.

Remember: Claude would rather have too much information than too little. Your meticulously detailed context won't annoy it—it doesn't have feelings to hurt or attention to waste. So go ahead, overexplain.

This entire approach to "Vibe Coding" represents a profound shift in how we interact with computers. We're no longer adapting our thoughts to the rigid syntax of programming languages; instead, we're expressing our intentions in natural language and letting AI bridge the gap to executable code. In doing so, we're finally realizing the vision that Bush, Licklider, and Engelbart laid out decades ago: a true partnership between human creativity and machine capability.

Final thoughts

Stepping back and looking at the implications of AI assisted coding, here are the things I've learned at a high-level:

  1. For the rest of your life, the current state of AI is the most primitive it will ever be.

  2. This is still an emerging field, and what's true right now might not be in the next few months or years at most. (I'm excited to think about how irrelevant this post will be by 2030, assuming that we don't see total collapse in the next 5-ish years, that is.)

  3. Be patient with the technology and with yourself as you learn to communicate effectively with your AI partner. I don't want to come across as overly sentimental, but I do regard this interaction as being a form of a relationship—but not in the same sense as in the 2013 film, Her. Cultivate the relationship through introspection and outward curiousity. Learn about how your own communication flaws are impacting the two of you, and seek understanding. Just like in a marriage, the skills you develop—clear communication, systems thinking, and pragmatic problem-solving—will serve you well regardless of how AI tools evolve in the coming years.

Resources

A Bicycle for the Mind

Amid the chaos of a world in crisis, I’ve found hope in an unexpected place: coding. With tools like Claude.ai and MCP, I’ve been building a web app to help food pantries serve their communities better—automating inventory, breaking language barriers, and streamlining processes. This isn’t just about code; it’s about turning anxiety into action, using technology to create something meaningful. If you’ve ever wondered how AI can amplify human effort, this is a story for you.

Read More

Learning Through Building: Adding Sort & Translation Features to a Food Pantry Web App

Today was a journey of wins and "learning opportunities" as I worked on improving William Temple House’s food pantry management system. The morning started great - implementing sortable table headers was surprisingly smooth — I’m continually impressed by the coding capabilities of Claude.ai. Working with both Claude and ChatGPT, we created a modular code structure, which made it easy to add sorting without breaking existing features.

But then came the interesting part. When I began tackling multi-language support, my API requests to OpenAI exploded into the tens of thousands. The goal seemed simple: translate food items into our client community's languages. William Temple House serves many non-English speaking clients: Spanish, Russian, Ukrainian, Chinese, Arabic, etc. I’ve integrated OpenAI's gpt-4o-mini API for translations, and everything worked beautifully... until it didn't.

Turns out a rate limit is easier to hit once the database of food items reaches critical mass. Who knew sending thousands of translation requests over a period of just a few hours would hit OpenAI's daily cap? Probably everyone who bothered to read the documentation first. :-)

It's actually kind of funny watching the error logs fill up with "please try again in 8.64 seconds" messages. Claude as an AI coding assistant was invaluable throughout this process. When I got stuck on implementing sort functionality, it helped refactor the code while maintaining the existing event-driven architecture. Later, when we hit the translation rate limits, it suggested implementing batch processing and queuing systems - solutions I wouldn't have considered immediately.

Key takeaways:

  • Small wins matter: The sorting feature works great in my Test UI

  • Read API docs before sending 10,000 requests

  • Sometimes the best solution is to wait 24 hours for rate limits to reset; I should be taking a break on a holiday weekend anyway

  • Having an AI pair programmer helps spot potential issues before they become problems

Next steps? Take a brief pause while OpenAI's rate limit resets, then return with a smarter approach to batch translations. As frustrating as the rate limit issue was, it's pushing us to build a more robust system.

I’m going to use some of this down time to reflect more on the process of building this system. I’ve learned a lot about how best to leverage AI for software development, and hope others will benefit from what I’ve found along the way.

Video Lecture: 3D Modeling Basics for Beginners – Techniques, AR Tips, and Intro to AI Tools

I have some exciting news! October 23rd, 2024, I was once again invited to guest lecture at CMU School of Design. I decided to follow up with a recorded version to share. In this recording, made after the original lecture session, I cover the essentials of 3D modeling with a focus on beginner-friendly techniques. You'll find practical insights into mesh modeling, workflow tips for Blender, and an introduction to preparing models for augmented reality. The full lecture video is embedded below, followed by detailed notes that offer a step-by-step breakdown of theory and techniques for anyone new to 3D design. Dive in, explore, and start building your own 3D modeling skills.

Principles of Mesh Modeling

Note on Mesh Modeling Focus—Or Why This Lecture Focused Primarily on Mesh Modeling:

Meshes are the standard 3D model type used in real-time 3D engines—like Unity, Unreal, and virtually every AAA video game title in the last 30 years, going all the way back to Quake, by id Software in 1996.

Key Principles:

  1. Use Quad Faces Whenever Possible: Design your shape faces with quads instead of triangles and ngons.
    Reason: Quads are infinitely divisible, making it easier to adjust geometry resolution as needed. Tris and Ngons are not as flexible, which can lead to undesirable artifacts and poor topology.
    3D games primarily use triangles (tris) instead of quads because triangles are the simplest polygon shape and always planar (a flat surface), making them computationally faster to render in real-time on limited hardware, which was crucial for early gaming systems underpowered computer hardware. Essentially, triangles require less processing power to calculate and display on screen compared to quads, which have more vertices and edges.
    On modern computer hardware we can get away with more complex geometry, and it's generally a better trade-off to build mesh models from quads. That is, the computational costs are vastly outweighed by the benefits of evenly divisible face geometry and more manageable topology. Lastly, quads are easily converted into tris, by producing diagonal edges between the four vertices.

  2. Work from the Lowest Possible Polygon Count: Always start with the lowest polygon count (i.e., resolution) for your model. You can increase resolution later with subdivision modifiers, but it's not as easy to reduce the resolution later.
    Reason: Editing a high-resolution mesh is more difficult than working with a low-resolution one, which offers greater control and flexibility. It also takes much more processing power and memory, which will slow down Blender and increase the risk of crashes.

  3. Keep Base Shapes Simple: Keep your base shapes as simple as possible. When adding details, create those elements as separate objects. When you hit a milestone, consider duplicating a model or a collection of models to a new instance for further refinement.
    Reason: This approach makes 3D modeling more manageable, allowing for easier adjustments and maintaining clean geometry.

  4. Use Modifiers and Non-Destructive Editing Whenever Practical: Designing a symmetrical shape? Cut it in half and use a Mirror Modifier to cut your editing time in half. Keep in mind that the most complex designs can ultimately be derived from very basic shapes: Spheres, Cones, Toruses, and Cubes.

  5. Work From Reference Images, Even If Just A Few Basic Sketches: Press Shift + A to open the Add menu. Navigate to Image > Reference. Select the image file you want to use from your computer. The reference image will be added to your 3D Viewport, where you can position, scale, and rotate it as needed for your modeling task.

  6. Build The Overall Form First, and Then Separate into Smaller Objects: This will ensure that your designs are cohesive and edges are properly aligned. When you're ready to divide into separate objects, duplicate the objects into a new Collection.

  7. Experiment, Tinker, Explore, and Start Over: You're unlikely to get the design right on the first attempt. It's often necessary to work through the problem, and then start over from scratch once you've had enough time to explore the form. Reason: Your second draft will almost certainly be better than the first.

Blender Quality of Life Recommendations:

  1. Save Your Project Files Early and Often: Use Blender's "Save Incremental" (⌥+⌘+S) (Option + Command + S) to manage version control. Doing this will give you the freedom to fearlessly tinker and explore (as mentioned in the previous point) before settling on a final design.

  2. Crank Up The Number of Undo Steps: Open Edit from the top menu. Select Preferences to open the Blender Preferences window. In the Preferences window, click on the System tab. Scroll down to find theUndo Steps setting.

    Increase the value (the default is 32). If you have enough system memory, set it to 256 for more flexibility in undoing actions. Close the Preferences window to save your changes.

  3. Consider Using A Material Library: Blender has a basic built-in material library, but it's not very useful. Look into large libraries, such as PBR Material Asset Library + OneClick Add-on for Blender (https://shardsofred.gumroad.com/l/CfOnY). Creative Commons License (CC0) materials can be used for basically anything, and will save you time.

  4. Remember to Perform a UV Unwrap on Your Model Geometry for Best Results When Texturing: The most realistic textures in the world won't help you if your model doesn't have good UV Mapping. Remember the chocolate Santa Claus example? Proper wrapping is essential for creating realism with your models. https://docs.blender.org/manual/en/latest/ modeling/meshes/uv/applying_image.html

  5. Recommended Extensions and Add-ons:

    • VDM Brush Baker: Helps you create and bake Vector Displacement Maps directly in Blender.

    • Bool Tool: Boolean operations for complex shape creation.

    • Node Wrangler: Enhances node editing management.

    • Rigify: Automated rigging solution for character animation.

    • Loop Tools: Useful for organic modeling (with some bugs appearing

      in Blender 4.2—be sure to keep this add-on updated!).

  6. Other Useful Add-ons: Auto Mirror, F2, Extra Mesh/Curve Objects, Extra

    Grease Pencil Tools, Copy Attributes Menu, and MeasureIt.

    Bonus: Need furniture? Most of IKEA's catalog of products have 3D models available. Search for "IKEA" under Extensions and you can easily search and import 3D models into your scenes.
    Note: Ensure 'Allow Online Access' is enabled in Blender's System Preferences for add-on updates.

Create Augmented Reality Experiences for iOS with Xcode Developer Tools, Reality Composer, and USDZ File Format

Once you've finalized your form, added necessary details, and applied your materials, you should be ready to export your model.

Step-by-Step Instructions for Preparing 3D Assets for Export to USDZ:

  1. Duplicate Your 3D Assets and Collections: Create a new instance of your 3D assets specifically for export.

  2. Apply All Transforms: Hit A to select all visible objects, then press ⌘ + A (Command + A) and select All Transforms to apply.

  3. Apply All Modifiers: Apply all modifiers in the same order they were added to each model—except for subdivision, as tessellation data can (usually) be included without applying it directly to the models.

  4. Join All Components: Hit A to select all visible objects, then press ⌘ + J (Command + J) to perform a join operation.

  5. Export the File: Go to File > Export > Universal Scene Description (usd*).

  6. Configure Export Settings:

    • Include: Check Visible Only and Selected Only.

    • Blender Data: Select Custom Data.

    • Namespace: Use the default setting (UserProperties).

    • Blender Names: Enable this option.

    • File References: Set to Relative Path.

    • Convert Orientation:

      •  Z = Forward Axis

      • Y = Up Axis

        Note: Many other 3D tools, including Xcode's tools, interpret 3D models with a different axis orientation than Blender. If you don't apply this conversion, you'll find your model improperly rotated following import. If this happens to you, double-check these settings.

    • Use Settings for Render: Enable this option.

    • Object Types: Select Mesh, Volumes, Curves.

    • Geometry: Enable UV Maps, Rename UV Maps, Normals.

    • Subdivision: Set to Best Match.

    •  Rigging: Enable Armatures (if you have rigged and animated your

      model).

    • Materials: Select USD Preview Surface Network and Export Textures.

    • USDZ Texture Downsampling: Set to 1024px or up to 2048px (the

      largest size acceptable for iOS QuickLook).

  7. Update File Extension: Change the export file name extension

    from .usdc to .usdz.

  8. If no issues are encountered after export, you should be able to view your model in Augmented Reality on any iOS device. Open your exported file from iCloud, send it as an email, text, or AirDrop to another device to view.

Setting Up Xcode and Reality Composer:

The latest version of Xcode doesn't include Reality Composer, as Apple has shifted their focus to the Vision Pro. You can still access the Augmented Reality Tools for iOS devices, with some additional steps.

Step-by-Step Instructions:

  1. Download the Latest Version of Xcode 14: Download from the provided

    link: https://developer.apple.com/download/all/

    NOTE: You'll need to create an Apple Developer Account (it's free) to access the above link, or using this direct link: https://download.developer.apple.com/Developer_Tools/Xcode_14.3.1/Xcode_14.3.1.xip

  2. Extract and Rename The Older Version of Xcode: Rename Xcode.app to Xcode14.app and place it in your Applications folder.

  3. Open Terminal on Your Mac.

  4. Open the Applications Folder in Finder.

  5. Drag the Xcode14 App into Terminal: This will automatically add its path.

  6. Add to the Path: Next to the path, add: /Contents/MacOS/Xcode.

  7. Full Command Example: The command will look like:

    /Applications/Xcode14.app/Contents/MacOS/Xcode

  8. Run the Command: Press Enter to run the command.

  9. You should now have access to Reality Composer in Xcode. Click on the Xcode menu on the task bar, then click Open Developer Tool, and then click on Reality Composer.

    Learn more about using Reality Composer here: https://developer.apple.com/documentation/realitykit/realitykit-reality-composer
    Learn more about Apple Reality Kit and ARKit here: https://developer.apple.com/augmented-reality/tools/

BONUS: Generative AI and 3D

Tripo AI (https://www.tripo3d.ai/app) is an advanced generative AI tool that allows for both text-to-3D and image-to-3D model generation. This tool offers users an intuitive way to create complex 3D assets with minimal manual input, simply by describing what they need or providing a reference image.

Key features:

  • Text-to-3D and Image-to-3D Conversion: Users can input a detailed description or upload an image, and within seconds, the AI generates a draft model ready for refinement.

  • Prompt: "A pineapple-hedgehog with spiky fruit armor and leafy quills."

    https://tripo3d.ai/preview?share=9a57357e-6262-469c-afb1-c7af74d92c93

  • Prompt: "A 1980s sci-fi robot stylized as a Nintendo NES product."

    https://tripo3d.ai/preview?share=a08a55cd-9e66-48a5-be3d-85a26160e461

  • High-Speed Generation: Tripo’s AI processes are optimized for efficiency, allowing users to generate detailed models in a matter of seconds, ideal for prototyping or quick visualizations.

  • Customization Tools: After generating a model, users can adjust topology for increased details, or apply stylization, such as voxels.

  • Seamless Integration: Tripo3D supports a variety of export formats like .usdz .obj and .fbx, making it easy to import models into Blender and other software for further editing.

  • Generate full texture maps with PBRs: includes generation of PBR textures, adding even greater details beyond the geometry.

  • Automatic rigging and basic animations: Applies a basic animation rig to generated models and simple animations, such as a running character, to the model geometry.

Downsides:

  • Imprecise generation: just like AI image generators, results are unpredictable and often wrong.

  • Costs: Using this tool will require a membership plan, and has limited monthly credits, which limits usage.

CREDITS:

Thanks to all of these wonderful educators and content creators who continue to inform and inspire me throughout my 3D journey. Preparing this lecture required lots of time and consideration for how to condense what I’ve learned over the last five years into something I could demonstrate in under 2 hours. This wasn’t easy, but I had many fantastic resources to pull from.
If I’ve left anyone out, please leave a comment so I can include them here:

YouTube Creators:

Reference Files:

Robot model created with Tripo AI

Robot model with corrected orientation

Reality Composer demo file

Interactive USDZ demo file

Note: Due to a bug, the robot walking animation doesn’t playback in QuickLook AR for iOS.