ð¨NSA Releases Guidance on Hybrid and Multi-Cloud Environmentsð¨ The National Security Agency (NSA) recently published an important Cybersecurity Information Sheet (CSI): "Account for Complexities Introduced by Hybrid Cloud and Multi-Cloud Environments." As organizations increasingly adopt hybrid and multi-cloud strategies to enhance flexibility and scalability, understanding the complexities of these environments is crucial for securing digital assets. This CSI provides a comprehensive overview of the unique challenges presented by hybrid and multi-cloud setups. Key Insights Include: ð ï¸ Operational Complexities: Addressing the knowledge and skill gaps that arise from managing diverse cloud environments and the potential for security gaps due to operational siloes. ð Network Protections: Implementing Zero Trust principles to minimize data flows and secure communications across cloud environments. ð Identity and Access Management (IAM): Ensuring robust identity management and access control across cloud platforms, adhering to the principle of least privilege. ð Logging and Monitoring: Centralizing log management for improved visibility and threat detection across hybrid and multi-cloud infrastructures. ð Disaster Recovery: Utilizing multi-cloud strategies to ensure redundancy and resilience, facilitating rapid recovery from outages or cyber incidents. ð Compliance: Applying policy as code to ensure uniform security and compliance practices across all cloud environments. The guide also emphasizes the strategic use of Infrastructure as Code (IaC) to streamline cloud deployments and the importance of continuous education to keep pace with evolving cloud technologies. As organizations navigate the complexities of hybrid and multi-cloud strategies, this CSI provides valuable insights into securing cloud infrastructures against the backdrop of increasing cyber threats. Embracing these practices not only fortifies defenses but also ensures a scalable, compliant, and efficient cloud ecosystem. Read NSA's full guidance here: https://lnkd.in/eFfCSq5R #cybersecurity #innovation #ZeroTrust #cloudcomputing #programming #future #bigdata #softwareengineering
Cloud Computing Solutions
Explore top LinkedIn content from expert professionals.
-
-
ð The Shift in Europe: Moving Away from US Hyperscalers ð©ï¸ As geopolitical concerns, data sovereignty, and pricing instability grow, European companies are making bold moves in their cloud strategiesâand the implications are massive. Over the past 15 years, reliance on public cloud giants like AWS, Microsoft, and Google has skyrocketed. But now, weâre seeing a strategic pivot unfolding across Europe, as organizations mitigate risks and embrace alternative solutions to protect their future. ð¯Â Why the shift? â  Data Sovereignty: Stricter data protection laws like GDPR and fears over compliance with laws like the US CLOUD Act are driving demand for European-managed cloud solutions and sovereign cloud providers. Organizations are prioritizing control over their sensitive data and leaning into platforms that support their unique privacy needs. â  Security and Trust: Concerns over potential government interference, espionage, and vendor lock-in are making European businesses rethink their current reliance on US-based hyperscalers. The rising interest in diverse, multi-cloud strategies and locally governed services reflects the growing importance of trust in cloud decisions. â  Economic Predictability: Increasing costs from hyperscalers have raised concerns about long-term pricing stability. Enterprises are recognizing that forward-looking cloud strategies need to include providers that prioritize pricing transparency and tailored solutions. ð¯Â Whatâs the result? A diverse and dynamic cloud ecosystem is emerging in Europe, leaning on open-source technologies, sovereign cloud providers, and tailored private cloud solutions. Platforms like OpenStack and others are paving the way for digital transformation without compromising on compliance or strategy. As businesses explore these new approaches, multi-cloud strategies, hybrid environments, and innovative pricing models are becoming essential for mitigating risks and staying competitive within an ever-evolving cloud landscape. ð¢Â This shift isnât just about technologyâitâs about geopolitics, trust, and long-term business resilience. Letâs embrace a future where diversity in cloud ecosystems fosters innovation, enhances security, and ensures sovereignty. What are your thoughts on this shift towards sovereign and multi-cloud solutions? ð Letâs discuss! #CloudComputing #DataSovereignty #SovereignCloud #MultiCloud #Geopolitics #Innovation
Why Europe Is Fleeing The Cloud
https://www.youtube.com/
-
Imagine you have 5 TB of data stored in Azure Data Lake Storage Gen2 â this data includes 500 million records and 100 columns, stored in a CSV format. Now, your business use case is simple: â Fetch data for 1 specific city out of 100 cities â Retrieve only 10 columns out of the 100 Assuming data is evenly distributed, that means: ð You only need 1% of the rows and 10% of the columns, ð¦ Which is ~0.1% of the entire dataset, or roughly 5 GB. Now letâs run a query using Azure Synapse Analytics - Serverless SQL Pool. 𧨠Worst Case: If you're querying the raw CSV file without compression or partitioning, Synapse will scan the entire 5 TB. ð¸ The cost is $5 per TB scanned, so you pay $25 for this query. Thatâs expensive for such a small slice of data! ð§ Now, letâs optimize: â Convert the data into Parquet format â a columnar storage file type ð This reduces your storage size to ~2 TB (or even less with Snappy compression) â Partition the data by city, so that each city has its own folder Now when you run the query: You're only scanning 1 partition (1 city) â ~20 GB You only need 10 columns out of 100 â 10% of 20 GB = 2 GB ð° Query cost? Just $0.01 ð¡ What did we apply? Column Pruning by using Parquet Row Pruning via Partitioning Compression to save storage and scan cost Thatâs 2500x cheaper than the original query! ð This is how knowing the internals of Azureâs big data services can drastically reduce cost and improve performance. #Azure #DataLake #AzureSynapse #BigData #DataEngineering #CloudOptimization #Parquet #Partitioning #CostSaving #ServerlessSQL
-
ð§ðµð² ðð®ð¹ð¹ð²ð± ð´ð®ð¿ð±ð²ð» ð°ð¿ð®ð°ð¸ð: ð¡ð®ð±ð²ð¹ð¹ð® ð¯ð²ðð ð ð¶ð°ð¿ð¼ðð¼ð³ðâð ðð¼ð½ð¶ð¹ð¼ðð ð®ð»ð± ðððð¿ð²âð ð»ð²ð ð ð®ð°ð ð¼ð» ðð®ð & ð ðð£ ð¶ð»ðð²ð¿ð¼ð½ð²ð¿ð®ð¯ð¶ð¹ð¶ðð Microsoft CEO Satya Nadella is redefining cloud competition by moving away from Azure's traditional "walled garden." The new strategy: supporting open protocols like Google DeepMind's Agent2Agent (A2A) and Anthropic's Multi-Cloud Platform (MCP), positioning Microsoft Azure Copilots and AI services for broad interoperability across cloud environments, including Amazon Web Services (AWS), Google Cloud, and private data centers. From my recent VentureBeat analysis, here are three reasons this shift matters: ð¡ Strategic Inflection Point: Microsoft is publicly endorsing and implementing A2A and MCP, aiming to make Azure a hub for genuine agent-to-agent interoperability across the industry. ð Enterprise Agility: By embracing open standards, Microsoft is reducing vendor lock-in and giving organizations greater freedom to innovate and manage AI workloads wherever they choose. âï¸ Technical Enablement: Azure's Copilots and AI platforms, such as Copilot Studio and Azure AI Foundry, are being built with open APIs and integration frameworks, simplifying and accelerating multi-cloud operations and adoption of interoperable AI solutions. ðð¼ððð¼ðº ð¹ð¶ð»ð²: The era of isolated clouds is coming to an end, and Microsoft is positioning itself as a key catalyst in that transformation. Full analysis linked in the first comment. #Azure #AI #MultiCloud #Interoperability #CloudStrategy #EnterpriseTech #Microsoft
-
Here's the last post sharing what I spoke about during PDP Week. Our moderator Christopher (2024 Global Vanguard Award for Asia) comes up with the most creative titles for panel discussions. He called this one 'Weather Forecast: Cloudy with a Chance of Breach'. Together with Aparna and Abhishek, we talked about privacy and security in the cloud. 1. Who do you typically engage with IRT privacy and security for the cloud? I wanted to dispel the misconception that if a company engages a cloud service provider (CSP) to store your data, they are responsible for privacy and security, and the company doesn't need to do anything. Generally, the cloud customer is still responsible for security in the cloud e.g. configuring user access to data, services that the customer uses. The CSP is responsible for security of the cloud e.g. physical protection of servers, patching flaws. This is known as "shared responsibility" between the CSP and cloud customer. The extent of each party's responsibilities depend on the deployment used e.g. SaaS, PaaS, IaaS. 2. Shared responsibility also applies within organisations e.g. - IT helps with technical implementation and maintenance of cloud services - IT security helps protect data from unauthorised access - Privacy, Legal, and Compliance provide guidance on compliance with laws, and ensure that contracts with CSPs and vendors include privacy and security clauses 3. What tools/processes are involved in privacy considerations for securing cloud use? They include a Privacy Impact Assessment when e.g. new cloud services are used to process sensitive data, when cloud use involves data transfers to various countries. Privacy management tools include encryption, anonymisation, pseudonymisation, access controls. CSPs usually make audit reports available to prospective and current customers, you can request for them. Also, have a well defined incident response plan. 4. How do you implement and manage breach or incident response for the multi-cloud? Multi-cloud environments can be challenging, because each CSP may have its own set of interfaces, tools, processes for incident response. You need to develop a unified incident response framework that can be applied across all cloud providers, which defines standard procedures for detecting, reporting, and responding to incidents, and which can enable collaboration between different cloud environments. The framework must facilitate internal coordination between various teams, as well as external coordination with CSPs. CSPs play a critical role in incident response, as they control the infrastructure and have visibility into their own environments. Ensure that roles and responsibilities are clearly defined, that you understand your legal obligations IRT breach notification e.g. who you need to notify and by when. Get corp comms' help with communication strategies vis-a-vis affected parties, regulators, staff, and other stakeholders. #APF24
-
ðªðµð ð®ð» ðð ðð¶ð¹ð¹ð¶ð¼ð» ððð¿ð¼ ðð»ðð²ðð ð¶ð» ððµð² ð¦ð½ð¿ð²ð²ðð®ð¹ð± ð¦ðµð¼ðð¹ð± ð ð®ððð²ð¿ ðð¼ ððð²ð¿ð ððð¿ð¼ð½ð²ð®ð» ðððð¶ð»ð²ðð ðð²ð®ð±ð²ð¿â I was in Munich last week for the ServiceNow World Tour, and the enormous interest in the "Digital Sovereignity for Europe" breakout with Schwarz Group and STACKIT was palpable with people standing to even get to see the session. It's clear that #DigitalSovereignty has moved from a regulatory buzzword to a CEO-level strategic imperative. Now, with Schwarz Digits announcing a massive 11 Billion investment in a new AI and data center in #Lübbenau, Germany (where STACKIT will operate its 5th facility), the European tech landscape is ð§ðªð¯ð¢ðððº taking actions. Here's my take on the new dynamics and what decision-makers need to know: 1ï¸â£ ð§ðµð² ððð¿ð¼ð½ð²ð®ð» ðð¼ðð»ðð²ð¿-ðððð®ð°ð¸: The 11B investment is a direct challenge to US Hyperscalers. It's about more than just physical infrastructure. It's about building an independent, high-performance platform for AI and cloud that is governed entirely by EU law (GDPR-compliant, protected from the US CLOUD Act). This is about choice and control for European enterprises. I would say ð£ð¦ðµðµð¦ð³ ðð¢ðµð¦ð³ ðµð©ð¢ð¯ ð¯ð¦ð·ð¦ð³! 2ï¸â£ ð§ðµð² ððð½ð²ð¿ðð°ð®ð¹ð²ð¿ ð¦ðð¿ð®ðð²ð´ð ð£ð¶ðð¼ð: Major US software companies are adapting to keep a multi-trillion-dollar market. The old way was 'one cloud fits all.' The new alliance model is 'sovereignty-by-design.' They are now considering to partner with trusted European infrastructure providers like #STACKIT (part of Schwarz Digits) to offer Sovereign Cloud solutions. New alliances are forming! 3ï¸â£ ð§ðµð² ð¦ð²ð¿ðð¶ð°ð²ð¡ð¼ð ðð¹ðð²ð½ð¿ð¶ð»ð: This is where the rubber meets the road. The partnership between ðð¦ð³ð·ðªð¤ð¦ðð°ð¸ ð°ð¯ ððµð¢ð¤ð¬ðð is a prime example. It allows businesses to leverage the power of ServiceNow's AI platform (with e.g. full feature parity) while ensuring all data is hosted and processed securely within the StackIT cloud, meeting the European data and compliance requirements. It's a pragmatic path to both innovation and sovereignty. ðð½ That clearly shows that digital sovereignty is not about closing the door! It's about building our own foundation. The combination of local investment, strategic alliances, and platforms like ServiceNow on StackIT is creating a resilient and competitive digital future for Europe. What is your organization doing to secure its digital future while maintaining its sovereignty?
-
ETL vs ELT in Data Engineering ETL: Extract, Transform, Load ETL is the traditional approach: 1. Extract:Â â³ Data is extracted from source systems. 2. Transform:Â â³ Extracted data is transformed (cleaned, formatted, etc.) in a staging area. 3. Load:Â â³ Transformed data is loaded into the target system (usually a data warehouse). Â Pros of ETL: - Data is cleaned and transformed before loading, ensuring high-quality data in the target system. - Reduces storage requirements in the target system as only relevant data is loaded. - Better for complex transformations that require significant processing power. - Ideal for systems with limited computing resources at the destination. Â Cons of ETL: - Can be slower due to transformation before loading. - May require more processing power in the intermediate stage. - Less flexible if transformation requirements change frequently. Â Use Cases for ETL: - Working with legacy systems that require specific data formats. - Data quality is a critical concern and needs to be addressed before loading. - Target system has limited computing resources. ELT: Extract, Load, Transform ELT is a more modern approach: 1. Extract:Â â³ Data is extracted from source systems. 2. Load:Â â³ The raw data is loaded directly into the target system. 3. Transform:Â â³ Data is transformed within the target system as needed. Â Pros of ELT: - Faster initial load of data as there's no transformation before loading. - More flexible, allowing for transformations to be modified without reloading data. - Takes advantage of the target system's power - Raw data is preserved, allowing for different transformations as needs change. Â Cons of ELT: - More storage is required in the target system as all raw data is loaded. - May result in lower-quality data in the target system if not managed - Can be more complex to implement and manage. Â Use Cases for ELT: - Working with cloud-based data warehouses. - Flexibility is needed for transformations on the same dataset. - Target system has significant computing resources. Real-World Example: Customer Analytics Platform on AWS Consider a real-world scenario where a retail org wants to build customer analytics platform using AWS Â ETL Architecture: 1. Extract:Â Â Â - Use AWS DMS to extract data from on-premises DB. Â Â - Use Glue crawlers to catalog data from S3 containing log files and other semi-structured data. 2. Transform: Â Â - Use AWS Glue ETL jobs to transform the data. 3. Load: Â Â - Load the transformed data into Redshift, a DWH optimized for analytics. 4. Orchestration: Â Â - Use AWS Step Functions to orchestrate the entire ETL pipeline. Â ELT Architecture: 1. Extract: Â Â - DMS and Glue crawlers for data extraction. 2. Load: Â Â - Load raw data into S3 data lake. 3. Transform: Â Â - Use Athena for on-demand SQL transformations. Â Â - Use Redshift Spectrum to query both structured data in Redshift and unstructured data in S3. 4. Orchestration: Â Â - Use AWS Glue to manage the ELT process.
-
I'm always on the lookout for "AWS" scale customer case studies ð !! This recent blog about how Ancestry tackled one of the most impressive data engineering challenges I've seen recently - optimizing a 100-billion-row Apache Iceberg table that processes 7 million changes every hour. The scale alone is staggering, but what's more impressive is their 75% cost reduction achievement. ðð¡ð ððð-ðð¨ð°ðð«ðð ðð¨ð¥ð®ðð¢ð¨ð§ Their architecture combines Amazon EMR on EC2 for Spark processing, Amazon S3 for data lake storage, and AWS Glue Catalog for metadata management. This replaced a fragmented ecosystem where teams were independently accessing data through direct service calls and Kafka subscriptions, creating unnecessary duplication and system load. ðð¡ð² ðððððð«ð ðððð ðð¡ð ðð¢ðððð«ðð§ðð Apache Iceberg's ACID transactions, schema evolution, and partition evolution capabilities proved essential at this scale. The team implemented merge-on-read strategy and Storage-Partitioned Joins to eliminate expensive shuffle operations, while custom partitioning on hint status and type dramatically reduced data scanning during queries. ðð§ððð«ð©ð«ð¢ð¬ð-ðððð¥ð ððð¬ð®ð¥ðð¬ This solution now serves diverse analytical workloads - from data scientists training recommendation models to geneticists developing population studies - all from a single source of truth. It demonstrates how modern table formats combined with AWS managed services can handle unprecedented data scale while maintaining performance and controlling costs. More details in the blog at https://lnkd.in/gN-mvdUE #bigdata #iceberg #aws #ancestry #analytics #scale #apache