<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Microsoft's AI Team Azure Misstep Exposes 38TB of Data: Cloud Security Concerns for CIOs in Cloud Security</title>
    <link>https://community.isc2.org/t5/Cloud-Security/Microsoft-s-AI-Team-Azure-Misstep-Exposes-38TB-of-Data-Cloud/m-p/67933#M395</link>
    <description>We store the secret,&lt;BR /&gt;On my PC in V-PAM,&lt;BR /&gt;A made-up sec tool.</description>
    <pubDate>Tue, 05 Mar 2024 23:58:30 GMT</pubDate>
    <dc:creator>Early_Adopter</dc:creator>
    <dc:date>2024-03-05T23:58:30Z</dc:date>
    <item>
      <title>Microsoft's AI Team Azure Misstep Exposes 38TB of Data: Cloud Security Concerns for CIOs</title>
      <link>https://community.isc2.org/t5/Cloud-Security/Microsoft-s-AI-Team-Azure-Misstep-Exposes-38TB-of-Data-Cloud/m-p/67930#M394</link>
      <description>&lt;P&gt;Hi All&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;According to Wiz, the affected data included full backups of two employees' computers that held passwords to Microsoft services, secret keys, and more than 30,000 internal Microsoft Teams messages from 359 Microsoft employees. Wiz says that Microsoft's AI team's upload of training data containing open-source code and AI models allowed GitHub users to access the models on Microsoft's Azure cloud service. The files were accessed by an Azure feature called SAS tokens, which allow you to see data from Azure Storage accounts.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;A href="https://bnnbreaking.com/tech/microsofts-ai-team-azure-misstep-exposes-38tb-of-data-cloud-security-concerns-for-cios" target="_blank"&gt;https://bnnbreaking.com/tech/microsofts-ai-team-azure-misstep-exposes-38tb-of-data-cloud-security-concerns-for-cios&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Well if they can do it once, it will not be the last time they will do this.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Regards&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Caute_Cautim&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 05 Mar 2024 21:53:31 GMT</pubDate>
      <guid>https://community.isc2.org/t5/Cloud-Security/Microsoft-s-AI-Team-Azure-Misstep-Exposes-38TB-of-Data-Cloud/m-p/67930#M394</guid>
      <dc:creator>Caute_cautim</dc:creator>
      <dc:date>2024-03-05T21:53:31Z</dc:date>
    </item>
    <item>
      <title>Re: Microsoft's AI Team Azure Misstep Exposes 38TB of Data: Cloud Security Concerns for CIOs</title>
      <link>https://community.isc2.org/t5/Cloud-Security/Microsoft-s-AI-Team-Azure-Misstep-Exposes-38TB-of-Data-Cloud/m-p/67933#M395</link>
      <description>We store the secret,&lt;BR /&gt;On my PC in V-PAM,&lt;BR /&gt;A made-up sec tool.</description>
      <pubDate>Tue, 05 Mar 2024 23:58:30 GMT</pubDate>
      <guid>https://community.isc2.org/t5/Cloud-Security/Microsoft-s-AI-Team-Azure-Misstep-Exposes-38TB-of-Data-Cloud/m-p/67933#M395</guid>
      <dc:creator>Early_Adopter</dc:creator>
      <dc:date>2024-03-05T23:58:30Z</dc:date>
    </item>
  </channel>
</rss>

