Just fixing thin content wasn't enough for us at PlayAbly. We dug into Search Console with structured data and found what Google was actually rejecting. After we added schema and fixed some soft 404s and canonical issues, a batch of our pages returned to the index. Even a good sitemap can miss things if pages lack clear topical signals or schema. It's those details that matter most.
I keep seeing the same thing when I help media sites recover from being deindexed. Sites with tons of stale content and almost no solid backlinks struggle, regardless of their technical setup. My advice is to start with a backlink audit. Then, focus on getting a few quality links from trusted sites to your most important pages. This proves your value to Google much faster than just tweaking on-page content.
I've seen this on education sites before. A drop this drastic usually means there's a bigger issue than just thin content. We once had Google de-index hundreds of pages after a migration, and the real problem was slow server speed. Google just wasn't waiting for the pages to load. Check your crawl stats and test the load speed on a bunch of random URLs. If that looks okay, try submitting a few pages manually to see if Google re-indexes them before you fix everything.
We had this problem with Tutorbase where Google stopped indexing our pages. We tried everything with sitemaps and robot.txt, but that wasn't it. Google just needed clearer signals about what was important. We started adding more links between our articles and updating the popular ones with new info. Once people started clicking around on those key pages, Google noticed and came back to check them out more often. For your site, I'd try refreshing a couple of your best articles. It's a quick test to see what gets Google's attention again.
Nothing's more annoying than watching your articles vanish from Google when everything looks fine technically. I see this happen a lot with the sites I host. Usually it's server response time - Google hates slow sites. Check your uptime logs and core web vitals first. One client fixed their server lag and their indexed pages bounced right back. If that's not it, try resubmitting your best articles through Search Console to see if that wakes Google up.
Yeah, I've seen this happen after big content updates. What fixed it for us was manually checking batches of URLs in Search Console and adding structured data to every single article. When we did that at Superpower, our tech posts started showing up again and our crawl budget wasn't getting wasted. My advice is to get really specific with your schema and use the Inspection tool. The problem is rarely just technical, it's about whether Google thinks your content has any authority.
I've seen this de-indexing happen before. It usually comes down to missing structured data or pages that are too slow for Google to crawl. My team found that improving Core Web Vitals, like compressing images and cutting back on scripts, really helps with crawl rates. We always review the NewsArticle schema and resubmit sitemaps after making changes. I'd check your page speed and structured data first. That's your best shot at fixing it.
Adding NewsArticle schema to our healthcare website didn't do much at first. But after we made sure to include the author, date, and full article text every single time, Google started indexing way more content over a few weeks. You also have to watch your page speed. If articles load slowly, Google just won't bother crawling them. Honestly, you need to do both the schema markup and the speed fixes together. That's what actually works.
From running big content sites, I've seen missing or broken schema, like NewsArticle, tank your indexation. Just tweaking our sitemap didn't do much at first. But once we added detailed structured data and made sure pages loaded fast, things picked up. Google crawls fewer articles if they're slow, so run a speed audit. You need to get both the technical signals and the content right. That's where I'd put my focus.
News sites often publish fast content which makes it easy for Google to mark pages as low value. When articles do not offer fresh insights the crawler may see them as less useful. Adding more depth to major stories can lift the overall quality of the site. Clear headlines and simple summaries can also help guide search engines more effectively. Google also reviews the health of the site and assesses the strength of each section. If many parts have thin stories the crawler may reduce its attention. Removing older articles that no longer hold value can help build stronger signals. A steady plan for content growth and internal linking helps crawlers find pages worth bringing back.
Your misunderstanding is that Google is attributing the issue to an Accessibility problem when it's actually telling you that 90% of your total content did not meet the E-E-A-T threshold set by Quality Raters. Therefore, the first thing you must do is stop modifying your website and implement a Strategic Content Consolidation and Pruning Plan. This plan entails identifying low-value, duplicate, or unoriginal content and significantly revamping those articles or removing them entirely from the index (with a 410 or no-index command). Second, focus on strengthening the ranking of the remaining 100 articles by building a solid internal linking structure that directs all link equity to those high-quality pages. By providing your audience with a high concentration of valuable information, you signal to Google's algorithms that your site is an extremely authoritative resource, subsequently raising your indexation threshold.
From my side this kind of crash usually means it is not a sitemap or robots issue anymore. I read "crawled currently not indexed" as Google saying I can see it but I do not trust or need it. If a lot of articles feel thin repeated or too similar to other sites the whole domain loses index priority. Then I look for silent bloat that drags everything down. Tag pages author archives pagination and URL variants can create thousands of low value URLs that get crawled first. That dilutes the site and makes even good stories look less important. I also check for near duplicate posts on the same topic instead of one strong updated article. Next I would clean and refocus instead of tweaking more tech. I would pick a few hundred best posts and deepen them with real value and link them from strong category hubs. I would noindex or merge low value archive pages and thin updates so the crawl budget goes to real articles. In my experience that quality and pruning combo is what brings pages back.
I've been in your exact shoes with a news site that grew fast. The crawled but not indexed drop wasn't technical at all. Google stopped trusting the overall value of the site. What helped me was deleting or noindexing everything that wasn't strong. I removed more than 40 percent of old thin posts and grouped similar topics into longer evergreen guides. Internal links from the home page and category pages pushed authority to the ones that mattered. I waited two weeks. Pages started returning to the index slowly and rankings improved because the whole site looked higher quality.