Are Mortgage Lenders Ready for Fannie Mae’s New AI Guidelines?

By

Knowledge Coop

If you’re a mortgage loan officer (MLO) using AI, this shift affects you.

New guidance from Fannie Mae and Freddie Mac (effective in August, 120 days after the letter’s publication on April 8, 2026) is changing how AI is treated across the industry. What was a simple productivity tool is now something lenders are expected to actively monitor, document, and explain. And while that may sound like a compliance issue, it reaches directly into the daily work of loan officers.

AI is no longer just a tool, it’s part of the loan process.

That includes anything from CRM automations to marketing tools to AI-assisted emails. Lenders now need to track how these tools are used, ensure they don’t introduce bias or errors, and be able to justify their use at any time.

For loan officers, this adds a new layer of responsibility. AI can still help you move faster, but you’re accountable for what it produces. An AI-written email or message still needs to meet the same compliance and accuracy standards as if you wrote it yourself.

It also means being more intentional about the tools you use. Relying on unapproved AI tools, even for simple tasks, can create risk. As companies build internal oversight, you’ll likely be expected to understand and disclose what tools are part of your workflow.

While that may sound restrictive, it creates a clear opportunity. Most loan officers will either avoid AI out of caution or use it without fully understanding the risks. The ones who learn how to use it properly, within compliant, approved frameworks, will have a clear advantage.

What Loan Services Should Know Today:

  • Ownership of AI output: You are responsible for anything AI produces on your behalf including emails, texts, marketing regardless of how it was generated.
  • Compliance in communication: AI-assisted messages must still meet fair lending, advertising, and accuracy standards.
  • Tool awareness: You need to know what AI tools you’re using and how they fit into your workflow.
  • Use of approved tools only: Using unapproved AI tools can create compliance and data security risks.
  • Transparency with your company: You may be required to disclose which AI tools you use and how you use them.
  • Avoiding bias and errors: You play a role in reviewing AI outputs to ensure they are fair, accurate, and appropriate.
  • Data responsibility: Any borrower information used with AI tools must be handled securely and within company guidelines.
  • Human oversight: AI should assist your work, not replace your judgment or decision-making.
  • Staying trained and informed: You’ll be expected to follow company policies and keep up with evolving AI guidelines.
  • Audit readiness: Your use of AI may need to be explained or justified if reviewed by your company or regulators.

Fannie Mae’s new AI guidelines make one thing clear: servicing teams must move from informal AI use to structured, accountable, and consciously monitored operations. Those who act now will not only reduce risk but position themselves to use AI more confidently and competitively.

Recommended Reads