In a significant breach of data security, a team of white hat hackers uncovered a substantial leak involving 38TB of Microsoft’s internal information. The breach, associated with the tech giant's AI research division, arose from a vulnerability in the Azure Storage system.
The white hat hackers, originating from the cloud security firm Wiz, first identified the issue on June 22, 2023. They discovered a shareable link based on Azure Statistical Analysis System (SAS) tokens that allowed them access to sensitive information. "This leak was a classic example of misconfigured storage containers, which we know to be common exploits in cloud security," said a representative from Wiz.
"This leak was a classic example of misconfigured storage containers, which we know to be common exploits in cloud security,"
The internal data became accessible after a Microsoft employee mistakenly shared a URL pointing to an internal Blob store within Azure. This repository contained critical AI datasets and open-source code related to image recognition, utilized by the company’s AI research division. "When we found the link, we didn’t just see code; we found access to backups of users’ workstations and internal communications," the Wiz source added.
"When we found the link, we didn’t just see code; we found access to backups of users’ workstations and internal communications,"

Upon realizing the breach, the Wiz team promptly reported their findings to the Microsoft Security Response Center. Microsoft acted quickly to patch the vulnerability, rendering the SAS token invalid by June 24, and subsequently replaced the token on the GitHub page where it was originally published by July 7.
The nature of SAS tokens allows for potentially dangerous configurations; they do not expire by default. This feature makes them unsuitable for sharing sensitive information. A blog post from Microsoft dated September 7 highlighted, "Attackers may create a high-privileged SAS token with long expiry to preserve valid credentials for a long period." This incident serves as a stark reminder of the importance of securely managing access tokens.
Although the magnitude of the data leak was alarming, Microsoft reiterated that no customer data was compromised during the incident. "The data that was exposed did not contain any customer information, and the integrity of other Microsoft services was not at risk due to this AI dataset," stated a Microsoft spokesperson.
"The data that was exposed did not contain any customer information, and the integrity of other Microsoft services was not at risk due to this AI dataset,"
This incident not only raises concerns about Microsoft’s internal data handling procedures but also serves as a lesson for organizations everywhere about the potential pitfalls of cloud storage. "Businesses can never be too cautious with their data management policies. Misconfigurations can lead to major vulnerabilities," remarked a cybersecurity analyst.

Looking Ahead
With cyber threats on the rise, companies must prioritize safeguarding their data against external breaches. This leak particularly underscores the necessity for stringent data-sharing policies, regular audits of storage configurations, and embracing secure coding practices. Doing so could save organizations from dire repercussions in the future.
As the digital landscape continues to evolve, incidents like this one serve as important learning experiences. Organizations are reminded that vulnerabilities can arise from seemingly minor oversights, reinforcing the need for vigilance in data security practices. Moving forward, companies, especially tech giants like Microsoft, must assess their cloud security measures to prevent similar occurrences from taking place, ensuring that their data—and the data of their clients—remains secure.

