Why Zillow Listings All Sound the Same (And Why It's Costing Agents Leads)
Every listing description tool on the market trained on the same MLS corpus. Here's what that costs you in dollars and days on market.
A few nights a week I pull up Zillow in a mid-size market and scroll through listings the way other people scroll through Netflix. I'm not house shopping. I find the writing interesting.
The problem is that the writing is almost always the same. After five listings, the openings blur. After ten, I stop reading the descriptions entirely and scroll straight to the photos. The houses are not the problem. Every listing sounds like every other listing.
In April 2026, I ran 53 of the most-viewed active single-family listings in Nashville through Montaic's listing grader. The average score was 4.6 out of 10. Nothing cracked 7. Thirty-one listings scored a 3 or a 4.
The vocabulary is where the pattern lives. "Beautiful" appeared in 38% of the listings. "Spacious" appeared in 34%. "Stunning" in 21%. "Boasts" and "move-in ready" each in 15%. The grader's lowest category score across the whole sample was cliche avoidance, at 4.2 out of 10.
The grader is not reading the listings and getting bored. It is reading them and writing the same feedback every time: "overloaded with generic cliches," "no emotional appeal," "no actual dimensions." The top complaint was the same complaint, over and over, in slightly different words. Which is exactly what the listings sound like.
This is not an agent problem. Agents are not writing bad copy on purpose. The reason every listing sounds the same is that every tool that writes listings, from ChatGPT to the dedicated real estate description generators, was trained on the same averaged MLS writing. When the training data is averaged, the output is averaged. When the output is averaged, your listing sounds like everyone else's. And when your listing sounds like everyone else's, you pay for it in two ways: dollars you leave on the sale price, and days you leave on the market.
Zillow's own research on 2.8 million listings
In 2016, Zillow Digs ran an analysis on 2.8 million home sales between January 2014 and March 2016. They controlled for property size, age, location, and the year and quarter of sale. Then they looked at sixty specific keywords in the listing descriptions and measured what happened to sale price and days on market when those keywords appeared.
The results were not subtle. Zillow's press release on the study shows listings that mentioned "barn doors" sold for 13.4% more than expected and 57 days faster. Listings mentioning "shaker cabinets" sold for 9.6% more and 45 days faster. "Farmhouse sink" earned 7.9% more and 58 days faster. "Subway tile" earned 6.9% more and 63 days faster. "Quartz" earned 6% more and 50 days faster.
These are not marketing claims. This is Zillow's own research on Zillow's own listings, published on Zillow's own press site. The data has been public for ten years and nothing has meaningfully replaced it at scale.
Now put that next to what the default listing description actually says. The default opens with a greeting to the buyer. The default describes the house using one of four or five generic approval adjectives. The default closes by telling the reader not to miss this chance. These are the phrases that the Zillow research found do not move sale price or days on market at all. The entire category of listing description tools is trained to produce exactly the phrases that Zillow's own research already proved are commercially inert.
Why every listing description tool produces the same phrases
Three things are worth understanding before the rest of this piece makes sense.
Every listing description tool trained on the same corpus. ChatGPT trained on the public web, which includes millions of scraped MLS descriptions. The dedicated listing tools trained on their own scraped MLS data, which overlaps almost entirely with the same corpus ChatGPT used. There is no competitive data moat in the listing description category. It is all the same data.
Averaged training data produces averaged output. A 2025 paper on "Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM Diversity" identified the root cause. Annotators who rate language model output systematically favor familiar text over distinctive text, because familiar text reads as "correct." Models learn that familiar text gets rewarded, so models produce familiar text. The bias is in the data, not in the algorithm. You cannot prompt your way out of it.
Structured formats amplify the collapse. A May 2025 paper on "The Price of Format: Diversity Collapse in LLMs" found that when a model is trained to produce outputs in a standard template, which is exactly what an MLS description is, the format itself suppresses diversity further. The more standardized the template, the less distinctive the output. MLS is one of the most standardized templates in consumer-facing content. The format compounds the training data problem.
Put those three facts together and the picture is clear. Every tool was trained on the same data. The data rewards averaged phrasing. The MLS format compounds the averaging. The output is a category-wide convergence on a handful of generic openers and approval adjectives. Your listing is the mean.
What the category average actually costs you
Here is what the compound effect looks like in practice.
An agent opens ChatGPT, pastes "write me an MLS description for a three-bedroom two-bath in Nashville with a renovated kitchen," and gets back a paragraph that opens with a greeting and describes the kitchen as beautiful and renovated. The agent pastes it into the MLS. The MLS syndicates it to Zillow, Redfin, Realtor.com, and a dozen secondary portals. Now that listing is one of approximately half a million Zillow listings active at any given time, and it is indistinguishable from the other half million.
Meanwhile, a 2023 study from Iowa State University led by Cheng Nie and co-authored with researchers at Georgia State, the University of Illinois at Chicago, and Nanjing University looked at how listing descriptions actually influenced buying decisions. They found that "experience attributes," meaning descriptive language that paints a specific sensory picture, correlated with higher sale prices. Generic approval language did not. The researchers used examples like "upscale bathroom fixtures" and "sunlit kitchen" for experience attributes. The ChatGPT-generated description in the scenario above has zero experience attributes. It has only generic approval language.
This matters because online listings are the first-touch channel for almost every home buyer. The National Association of Realtors 2025 Home Buyers and Sellers Generational Trends Report, published April 2025, found that online listings are the most-used information source in the home search across every generation, and that 86% of all buyers used a real estate agent but started their search on a portal. If the listing loses attention on the portal, the agent loses the lead before the agent ever enters the picture.
The compounding is the business problem. Every generic-sounding listing trains the reader to scroll faster. Every faster scroll makes the next generic listing even more invisible. The agents who are still writing distinctive descriptions end up with disproportionate attention because they are the only ones left in the signal.
Five patterns in every tool-generated listing
Five patterns show up in almost every tool-written listing description. If your last five listings match more than two of these, you are on the wrong side of the Zillow research.
- Generic approval openers. Every tool-written listing opens with one of about six phrases. A greeting to the buyer, an adjective about the house, an appeal to the seller's pride. The Zillow keyword data shows that none of these phrases move the price or the days-on-market needle.
- Feature list without sensory context. A line of "3 bed, 2 bath, hardwood floors, granite counters, stainless appliances" is searchable but not readable. It does not tell the buyer what the house feels like. The Iowa State research specifically flagged that experience attributes matter more than feature lists for sale price.
- Target-demographic pandering. Phrases that tell the reader which kind of buyer should want the house. These are also the phrases that trip the Fair Housing analysis in our earlier piece on Fair Housing Act listing description rules. They are commercially useless and legally risky at the same time.
- Location cliches. Generic proximity claims that do not contain specific distances or named destinations. "Close to everything." "Minutes from downtown." The portal reader has read these phrases fifty times this week.
- Manufactured urgency. Urgency language that is not backed by specific scarcity information. The digital equivalent of a screaming car commercial. The reader notices it and scrolls past faster.
The common thread across all five is that these are phrases the tool produces because the tool's training data is full of them. The tool is not making a choice. It is reverting to its mean. Your listing is the mean.
Four steps to get out of the category average
Step 1: Audit your last five listings against the Zillow keyword list. Pull up your five most recent listings. Count how many of the high-performing keywords from Zillow's 2016 study actually appear in your descriptions. Then count how many of the five patterns above appear. If the ratio is worse than two low-performers to one high-performer, you are probably losing sale price and days on market to the category average.
Step 2: Write the opening sentence by hand, every time. The first ten words of a Zillow listing are doing 80% of the work. They decide whether the reader scrolls or stops. Even if you use a tool for the rest of the description, write the first sentence yourself, in your own voice, about the specific house. Specificity is the thing the tool cannot fake.
Step 3: Describe the experience, not the feature list. For every feature, write one sentence about what it feels like to use it. "Stainless appliances" becomes "the kitchen is bright enough to read a recipe without turning the light on." "Hardwood floors" becomes "the original red oak floors run the length of the first floor and have the patina you only get from 80 years of being walked on." This is what the Iowa State research meant by experience attributes.
Step 4: Use a tool that trained on your writing, not on the average writing. The underlying problem is the training data. If the tool was trained on the same corpus as every other tool, it will produce the same output. The only category-level fix is a tool that trained on your specific writing and your specific listings, so the averaging works in your favor instead of against you. This is what Montaic does, and it is the reason we built the product.
What Montaic does differently
I built Montaic because I spend too much time on Zillow. Not as an agent. As a reader who also runs a coffee shop and shoots photography and knows what it sounds like when the same phrase has been used 500 times that week.
The coffee shop teaches you what generic marketing sounds like. When you run a food business, you hear the same three or four hospitality phrases from every competitor who used the same shop-for-pastry marketing playbook you could have used, and you learn to hate those phrases because they do not describe anything. The photography teaches you what "a reason to click" looks like. A listing photo either gives the viewer a reason to stop scrolling or it does not. The portal browsing teaches you what category-average copy feels like from the reader's side, not from the writer's side.
Montaic is a listing description tool that trained on your writing, not on the average. Here is what that means in practice:
- Writing style lock. You paste in three listings you have written that you were proud of. Montaic trains on those as the voice sample. Every subsequent description uses your phrasing, your sentence length, your rhythm.
- Experience attribute scoring. Every draft is scored on how many specific sensory details it contains versus how many generic approval phrases. The score is visible before you publish.
- Fair Housing screening built in. Every draft is screened against the 24 C.F.R. § 100.75 risk list before it leaves the tool. No separate compliance step.
- Portal-aware formatting. Montaic knows which portal the listing is going to and adjusts sentence length and structure to match the scroll behavior on each.
Montaic does not replace your judgment about the house. It does not know what the kitchen smells like at seven in the morning, or which window gets the best afternoon light. Those are the things you have to bring. What Montaic does is make sure that when you write about those things, the writing sounds like you and not like every other tool in the category.
Frequently asked questions
Does listing description quality actually affect sale price, or is that just marketing advice?
It affects sale price, and the numbers are public. Zillow's 2016 analysis of 2.8 million home sales controlled for property size, age, and location, then looked at specific keywords in the listing description. Listings mentioning "barn doors" sold for 13.4% more than comparable homes. Listings mentioning "shaker cabinets" sold for 9.6% more. The effect was measured in dollars, not in engagement metrics. A peer-reviewed 2023 study from Iowa State University confirmed the same direction using different methodology and a different sample. The research has been consistent for a decade.
Is ChatGPT safe to use for MLS descriptions?
Technically yes, legally maybe, commercially no. Technically the tool will produce a description. Legally ChatGPT has no Fair Housing screening, and the agent is responsible for everything that gets published, so any violation is on the agent. Commercially, ChatGPT was trained on the same averaged MLS corpus as every other tool, so the output reverts to category-average phrasing that Zillow's own research shows does not move the price. Using ChatGPT for MLS descriptions gets you a legally exposed description that sounds exactly like your competitor's legally exposed description.
Why do all the listing description tools produce the same phrases?
Because they all trained on the same data. The training corpus for listing description tools is scraped MLS archives, which are heavily skewed toward the phrases agents already use. A 2025 paper on mode collapse in language models identified that annotators who rate model output prefer familiar phrasing, which makes models learn to produce familiar phrasing. The bias is in the data. No amount of prompt engineering can pull the model out of its averaged mean.
What counts as a good listing description under the Zillow research?
Specific. Sensory. Named features that are either high-end or distinctive. The Zillow keyword list favors phrases that describe actual materials and installations (shaker cabinets, subway tile, farmhouse sink, quartz counters) over generic approval phrases. A good description tells the reader what the house is, not how the writer feels about the house.
What should I do if I have been using ChatGPT or a generic listing tool for the last year?
Audit your five most recent listings. Count how many high-performing Zillow keywords appear versus how many of the five tool-average patterns appear. If the ratio is bad, the fix is not to prompt better. The fix is to switch to a tool trained on your own writing, so future descriptions sound like you instead of sounding like the average. Old listings do not need to be rewritten. New listings starting from now will separate you from the category default within a quarter.
Does Montaic work for agents outside residential real estate?
Montaic is built primarily for residential real estate. The writing style lock and the portal-aware formatting are tuned for Zillow, Redfin, and Realtor.com. We do have early customers in yacht brokerage, and the underlying approach (train on your writing, not on the average) works the same way there. If you are in a listing category other than residential real estate, reach out and we can talk about whether Montaic fits.
Closing
Every Zillow listing sounds the same because every listing tool trained on the same averaged MLS corpus, and the averaged corpus produces averaged output. Zillow's own research shows that the averaged output is leaving dollars and days on the table. The fix is not a better prompt. It is a tool that trained on your writing instead of on everyone else's.
If you want to see where your current listings land on the experience-attribute scale, run one of them through the Montaic grader for free. No account. The grader returns a score and a breakdown of which phrases are helping and which are hurting, and it takes about a minute.
More Resources