Page 1,774«..1020..1,7731,7741,7751,776..1,7801,790..»

Getting Started With Kubernetes at the Edge – Container Journal

Gartner estimates that only 10% of data is produced and handled outside of traditional data centers today. Because of the rapid spread of the internet of things (IoT) and increased computing power on embedded devices, this figure is expected to rise to 75% by 2025. McKinsey identifies over 100 possible edge computing use cases with a potential $200 billion in hardware value produced over the next five to seven years.

In this article, you will learn how Kubernetes is quickly becoming one of the most popular solutions used by businesses to incorporate edge computing. You will also learn about the benefits of edge computing, the specific benefits Kubernetes offers to assist with edge computing and how Kubernetes distributions could be used for edge computing.

Edge computing has received a lot of attention and has become somewhat of a buzz word, but what does it truly mean for a business? Lets look at some of the most significant advantages of edge computing.

For applications with vast volumes (and velocity and variety) of data, processing data at the edge may be more efficient than paying for the bandwidth necessary to process that data in the cloud. To lessen the strain on your own cloud servers, computation can be done on client devices such as the users PC or even their smartphone in some circumstances.

You may also limit the quantity of data in the long run by doing real-time processing at the edge and just transmitting lower granularity to the cloud for long-term historical analysis.

Moving the computing resources closer to users reduces latency, giving them a better experience. Because fewer round trips to data centers result in lower latency and lower bandwidth costs, new functionality and features become available.

End users will be more reliant on a well-designed application that takes advantage of edge computing. Even if a network connection to data centers is lost, critical work may still be completed by using edge computing capabilities.

Edge computing also could assist your architecture in eliminating single points of failure.

Edge computing can increase the security of your software application as well as the privacy of your users. When compared to a more traditional design, storing more data at the edge and away from centralized data centers helps limit the blast radius of security breaches.

Edge computing can also make it simpler to comply with data privacy requirements. Instead of transferring data to the cloud and keeping it, it can be processed on the users own device or at the edge before being erased or altered to eliminate personally identifying information.

Now that youre aware of the numerous advantages of implementing edge computing, the question is how to go about doing so. There are several possible alternatives, ranging from developing your own platform to using a service supplied by another organization. Another approach for dealing with edge computing is to use Kubernetes.

From a technological and economic standpoint, there are various advantages to employing Kubernetes for edge computing. Kubernetes is already technically intended for operating across data centers and coping with difficulties that are inherent in edge computing. As a result, the transition from multi-region data centers to various edge locations isnt all that difficult.

From a commercial standpoint, by selecting Kubernetes as your edge computing platform, you gain the advantages of the enormous community which, over time, saves you time by preventing you from having to implement several common features and guarantees the project is maintained and safe.

Kubernetes Distribution Options

There are various choices for edge computing with Kubernetes in terms of both architecture and Kubernetes distribution. These distributions address some of the issues that make use of conventional Kubernetes for edge computing difficult.

KubeEdge is probably a suitable option for explicit separation of edge and cloud as well as an overall Kubernetes deployment. KubeEdge provides an edge environment on a cloud platform and connects it to the main Kubernetes deployment through an edge controller. This results in a setup that is identical to a conventional Kubernetes deployment via both the edge and the core. However, administering the edge component is simpler since it requires less detailed rule-building to effectively guide edge pods to edge nodes and construct backup pathways. To access edge elements, KubeEdge additionally contains a lightweight, edge-centric service mesh.

K3s, a Rancher-developed small-footprint Kubernetes distribution thats designed for edge missions with limited resources, is another package that may be crucial for Kubernetes at the edge. The footprint of K3s can be half or even less than that of the typical Kubernetes distribution, and it is fully CNCF-certified such that both are powered by the same YAML configuration files. By establishing an edge cluster, K3s further isolates the edge from the cloud. This configuration is advantageous in situations when edge pods cannot operate outside the edge due to limitations on resources or latency reasons. However, K3s features non-redundant components that might be risky, including database components like SQLite, and it can be more challenging to manage a distinct K3s edge cluster if administrators can assign the same pods to both the edge and the cloud.

Canonicals MicroK8s is a powerful, Cloud Native Computing Foundation-certified Kubernetes distribution. Below are some of the key reasons, why it has become a powerful enterprise computing platform:

Additionally, MicroK8s can coordinate fully fledged cloud resource pools while having a tiny enough footprint to operate in environments with limited resources. Thus, MicroK8s is undoubtedly the edge Kubernetes solution that is the most edge-agile, and it does it without requiring a complicated installation or operation.

The most important question to ask when running Kubernetes at the edge is whether your organizations edge resources are comparable to those in the cloud. If they are, the more effective setup is a standard Kubernetes deployment with set node affinities and related pod-assignment parameters to steer edge pods to edge nodes. For this kind of setup, consider KubeEdge if the edge and cloud environments are symbiotic rather than unified.

The more dissimilar the edge and cloud environments or requirements are, the more logical it is to separate the two, especially if edge resources are insufficient to run standard Kubernetes. Use K3s or MicroK8s if you want common orchestration of both edge and cloud workloads.

The book IoT Edge Computing With MicroK8s gives a hands-on approach to building, deploying and distributing production-ready Kubernetes on IoT and edge platforms. This edition has 400+ pages of real-world use cases, scenarios to help you successfully develop and run applications and mission-critical workloads using MicroK8s. Some of the key topics covered are:

By the end of this book, youll be able to use MicroK8 to build and implement scenarios for IoT and edge computing workloads in a production environment.

The key takeaway here should be the adaptability of Kubernetes for edge computing. Companies of many sizes and in many sectors are leveraging Kubernetes capabilities to improve the efficiency and reliability of their applications.

Related

See the original post:
Getting Started With Kubernetes at the Edge - Container Journal

Read More..

Exploring the Benefits of Containerization – Container Journal

As technology developments expand and become more complex, the software we rely on for our website redesigns and customer communications also increases in complexity. Thats exacerbated by the shift to developing cloud-native applications built using microservices and running on platforms like Kubernetes. It can make it tricky for developers to navigate when making changes, installing new features or testing applications, as the infrastructure behind programs is made of many different parts. Learning to manage and organize these can be tricky for developers.

As part of the effort to simplify application development processes, many developers are leveraging containerization. It is one way of grouping the infrastructure of applications, making it easier for developers to interact with and manage. Its not a new method and has been evolving for several years now, allowing developers to become more confident in using it. Containerization also has many benefits for developers.

When you create an application, there are configuration files, dependencies, and other computing resources needed to make it run. Containerization moves these to a portable, self-contained computing environment called a container. A container strategy does not rely on virtualized operating systems using resources to make the application run instead, they do this independently with any host operating system or computing environment.

Multiple containers can be used simultaneously with your operating system, depending on the resources your system has available. It can store data, build microservices, or test and deploy on a larger scale with web applications. As containers dont depend on a specific computing environment, if you need to swap, add to, or remove containers, you can do this without managing each file and resource needed for the application individually.

Image sourced from Datacenterknowledge.com

Protecting data on your computer requires the encryption of sensitive files and folders with an additional layer by using either file encryption or container encryption. Container encryption is the better choice of the two as it involves creating a secure virtual drive that is capable of storing many encrypted files at once. Container encryption removes security risks since you only need to remember one password to access each container.

The more you work with containerized applications, the more benefits youll notice in how they work and organize your application resources. Your experience with containers will somewhat depend on the application youve created and its functionshow a softphone works using containers will differ from those used for a calendar application. However, these are seven of the most common and helpful benefits provided when containerizing any application.

As containers include all the resources they need to function, they dont drain resources from the server or other containers being used. This allows the high performance of each container in its functions, even when using the same server for multiple containers. Should one container crash, this wont affect the functionality of your other containers on the same server. Similarly, the server will continue to work, enabling developers to continue using it.

Containers also allow quicker start-up times with your applications and running on your systems as all the resources needed are easily accessible. As they use the operating system of the machine they are installed on, they dont need their own to function, making them smaller in capacity. It again helps to improve the speed of your application, helping you to use its functions more efficiently and with fewer delays or lags.

A key feature of containers is their portability, meaning they dont rely on an operating system or specific server to function. Instead, they integrate with whatever server you choose, taking their code and resources along. It means once youve built your container, you can use it anywhere without having to adapt or rewrite parts to make it function. Regardless of the computing environment, the container will still work as the developer intended.

Accessing containers from any server or computing system simplifies sharing and distributing them, eliminating the need for changes depending on your operating system. This also makes containers for applications like contact center as a service (CCaaS) solutions easier to use as performing the functions is always the same across platforms. Instead of requiring a specific device or server to perform functions, you use them anywhere with any available system.

As you continue developing an application, using containers makes it quicker to distribute changes and new features across multiple servers and computing environments. It is particularly important with bug fixes and upgrades, as this prevents the damage that an issue can have on users by resolving it efficiently. Containers allow developers to deploy these changes on each server without having to rewrite the code or make adjustments.

Also, having all the resources for an application in one place makes them more accessible for developers to experiment with and create new features using app definition and build tools. As the files and coding needed are within the container, developers dont have to search for the appropriate resources or repeat their changes for use on multiple servers. Once the application feature is written using the container, this can be tested and shared without delay.

Depending on your application, you may need space to grow and further develop features. As each container is independent, these can easily be scaled up as your application requires. It could be through adding additional containers to your server for your CRM migration (services), increasing the functions available to you and your customers. Alternatively, you can move storage application containers to different servers to create more space.

Image sourced from Researchgate.net

Similarly, if you no longer need a function or resource, you can remove an irrelevant container to replace it with another more accommodative option. Using container orchestrators can help with this by managing the containers youre running, keeping only those youre using for the application open to improve speed and functionality. Also, having multiple isolated containers for the same application can make it more available for users by improving the uptime.

In the same way that one container on your server experiencing issues doesnt affect other containers, so too does the security of one container not compromise the security of the other containers. As each container is isolated from the others and can be removed from the server or use a cloud-first strategy, its easier to manage security breaches and the effect that they have. It reassures your application users that their data and app usage is safe.

Likewise, containerization can act preventatively to stop security issues. Part of the features the developer can create with the containers can include firewalls and malware protection. These prevent malicious software or cyber attacks from damaging your resources and protect your servers. Even if your server experiences security issues, the containers interact little with it and wont necessarily be impacted as they rely on their resources to maintain their application.

Traditional servers can be expensive and difficult to maintain with complex infrastructures and configurations. These can require downtime periods to allow for upgrades, costing you both financially and in user satisfaction. It isnt ideal for your application or business, as you want to have the best value for money. Instead, using containers means that your application doesnt have to use servers, allowing you to look for cheaper options.

Image sourced from Quora.com

Servers can also be susceptible to hacking and require manual upgrades. As containers dont rely on servers to function, they can use other available options to run your application. This gives you the choice of VoIP or analog and other solutions for your containers, such as cloud options or virtual machines. These can help set up automatic upgrades, reduce the risk of hacking, and make your application more reliable and consistent for users.

We know the importance of saving your work as you go along and the same applies to your applications. Creating data backups and recovery strategies can limit the damage caused by hackers, keeping necessary resources secure and available for developers to rebuild with. Using containers can make storing and updating backups easy, so theyre ready to be used. It also allows your developers to experiment and test new functions using the backups.

There are many backup options available, including the ability to duplicate your containers or use a versioning system. By making multiple versions of the same container, you can replace one container with another should it be tampered with. Alternatively, a versioning system stores the configuration of your application. With containers, you can take different backup approaches for each component according to what works best for storing your application information.

If your application infrastructure is getting out of hand, introducing containerization may be the solution. As an easy way to organize the components of your application, making it more portable and independent, containers can transform the upkeep of your app. Containerization benefits the developers who work directly with the resources and the users applying the functions for porting a phone number or organizing social media strategies.

Image sourced from Redhat.com

Containerization is a process and will take time and effort to transfer all the appropriate resources into containers for your applications. However, the results can reduce the time and effort needed to manage your application. Likewise, by using containers on new projects too, you can avoid organizational issues with the resources your applications require to perform. Start using containers for your applications and experience the benefits for yourself.

Related

Originally posted here:
Exploring the Benefits of Containerization - Container Journal

Read More..

Recommendations in creating a data center recovery plan – Malaya

IT may be funny but there is a common misunderstanding that data stored in the Cloud means data just floating around somewhere so it can be retrieved at any time. The truth is, data stored and managed in the Cloud requires data centers, huge facilities that run either magnetic, optical or digital storage devices to keep and back-up (at least twice) all the data we let evaporate into the Cloud.

With the data center becoming the repository of all our informationpersonal and professional data, photos and illustrations, spreadsheets and documents, music and movies. Keeping these important artifacts intact at a data center also requires that the data center be constantly humming.

Data redundancy, fire-proofing hardware, software back-ups, cybersecurity protocols and power outage management all help the data center system administrators in keeping customers data intact and accessible on a physical level. Creating business continuity solutions for digital critical infrastructure starts with a disaster recovery plan that eliminates a common cause of data center downtime: human error.

Power management solutions expert Vertiv recommends evaluating all risks that include natural catastrophes like hurricanes, flooding, tornadoes, fires, earthquakes, and volcanoes or even radiation exposure, hazardous waste, or explosives. Once the risks are identified and evaluated, creating a secure evacuation plan follows. Human safety always comes first and an escape-and-rescue strategy to evacuate employees who may be in danger must be in place.

From another perspective, it also advises weatherproofing the data center, with appropriate precautions to protect against storms, flooding, or other weather-related threats. Earthquakes are another calamity that data centers are built to withstand. However, Vertiv still advises checking if servers are safely stored and secure in their racks. Doors should be sealed against strong winds and rain, making every effort to prevent any water from entering the server rooms because water is the enemy of the data center.

Back up data and check the generator. Every week, several data centers perform routine data backups keeping in mind that one back up is no back-up at all. In adopting a daily backup routine also consider the location of the backup data. A safe off-site storage is recommended. There is a tendency to set and forget a generator, but that piece of machinery requires maintenance and upkeep to ensure it performs as expected when needed. Test the generator regularly and ahead of any anticipated weather events. Line up at least three vendors to deliver fuel in the event of an extended outage.

Communicating with utilities and contacting vendors assures that there is continuity or at least a plan to ensure having a phone and internet connection, water, or power. Establish early communication with utility providers to create backup plans. Make a contact list and prepare a communication strategy. Prioritize the vendors on your list who must be contacted in an emergency. By doing these two more focus can be put on more pressing demands during the crisis.

Plan emergency staffing and trust the team. Local workers might not be available to work in a significant crisis. They might have evacuated, had severe issues travel or shelter issues. Find ways to guarantee to have a staff on-site, including emergency services and setting up crisis housing close to the data center. Gathering IT, facilities, security, HR, communications, legal, logistics, and information security and briefing key people of their roles during the crisis. If the typical lines of communication are unavailable, have a plan for communicating with that team.

Recommendations for administrative and technical aspects include confirming insurance coverage to check if additional protection for the infrastructure may be necessary, as continuity of business insurance can make up for lost revenue if the data center is down for a week. Remember the edge, is sound advice to oversee edge sites that sprout out from the vastness of a dispersed network where an enterprise data center today is merely a component. Many companies oversee numerous edge sites.

Vertiv puts particular attention to minding the Cloud service provider as a lot of data and applications are stored there. These cloud servers are located in a data center, and knowledge of the Cloud providers protection and response systems in an emergency is both an assurance and lesson from which best practices can be taken from.

Consider the opportunists. Hackers consider catastrophes like natural disasters as a chance to access networks when peoples attention is diverted elsewhere. Ensure your information and physical security personnel are ready to deal with cyber threats.

The right data center services provider may also join the team and adequately assist with the risk assessment required to plan for a disaster adequately. These infrastructure specialists are specially qualified to guarantee a prompt and secure recovery.

With its expertise, Vertiv brings together cutting-edge solutions and services to ensure uninterrupted operations, optimal performance, and scalability of data centers, communication networks, and other critical IT facilities.

To learn more about how Vertiv supports the continuity of todays vital business, visitVertiv.com.

Visit link:
Recommendations in creating a data center recovery plan - Malaya

Read More..

The Worldwide Telecom Cloud Industry is Expected to Reach $103.6 Billion by 2030 – Benzinga

Dublin, Sept. 29, 2022 (GLOBE NEWSWIRE) -- The "Telecom Cloud Market Size, Share & Trends Analysis Report by Component (Solution, Services), by Deployment Type, by Service Model, by Application, by Enterprise Size, by Region, and Segment Forecasts, 2022-2030" report has been added to ResearchAndMarkets.com's offering.

The global telecom cloud market size is expected to reach USD 103.6 billion by 2030, according to this report. The market is anticipated to expand at a CAGR of 19.9% from 2022 to 2030. A telecom cloud is a next-generation network architecture that integrates cloud-native technologies, network function virtualization, and software-defined networking into a distributed computing network. Orchestration and automation are essential since the computing and network resources are scattered across clouds and locations.

Telco Cloud refers to shifting communications service providers (CSPs) from vertically integrating proprietary hardware-based infrastructure networks to cloud-based technologies. It is mainly used in the telecom business to refer to multi-cloud computing. The propelling drivers of the telecom industry are increased customer satisfaction, corporate agility, cost savings, and others. Also, the usage of standard computational hardware and automation reduces CapEx and OpEx resulting in increased adoption of telco cloud in the telecommunication industry.

It also delivers innovative bespoke B2B solutions, such as telcos may bring highly customized corporate products to market rapidly and affordably. Telco cloud makes it simple to collaborate with business service partners by providing access to public cloud services from any device, at any time. Additionally, it protects your consumers and profits from competitors; for instance, the telco cloud enables operators to swiftly alter business models to test new goods, services, and pricing schemes.

It also makes setting up new consumer experiences and communication channels easier. Furthermore, the lower CapEX and OPEX needs of telco cloud, better service resilience, and capacity to respond swiftly to faults and demand changes allow operators to maintain service levels and competitive pricing. These advantages result in lower client attrition.

The top trends in the telecom cloud industry are hybrid cloud hosting, Cloud Native Network Functions (CNNF), and telecom cloud collaboration. A hybrid cloud merges private and public clouds where the software and data are interoperable and portable. It allows telcos to optimize the operations with various patterns to manage workload. It improves resource allocation, optimizes infrastructure spending, provides enhanced organizational agility, and offers the ability to scale using the public cloud and controls available in the private cloud deployment.

Also, in the case of CNNF, Software-defined networking is replaced by NFV (Network Functions Virtualization), which provides more independence from proprietary servers and hardware. It provides a cloud-native architecture that combines VNFs and CNFs while adopting 5G features. This will provide maximum market coverage to telecom businesses looking to expand their services. Moreover, telecom cloud collaboration includes partnerships between hyperscalers and telcos which constitute a major cloud computing trend transforming the business.

Cloud service providers and telecom enterprises join forces to expand edge computing collaboration and 5G. Telecom cloud service providers are increasing their connectivity with the help of technology advancement to gain a competitive edge over their peers and capture a significant market share.

Telecom Cloud Market Report Highlights

Key Topics Covered:

Chapter 1. Methodology and Scope

Chapter 2. Executive Summary

Chapter 3. Market Variables, Trends, & Scope Outlook3.1. Market Segmentation3.2. Telecom Cloud Market Size & Growth Prospects3.3. Telecom Cloud Market - Value Chain Analysis3.4. Telecom Cloud Market Dynamics3.4.1. Market Driver Analysis3.4.2. Market Restraint Analysis3.4.3. Market Opportunity Analysis3.5. Telecom Cloud Penetration & Growth Prospects Mapping3.6. Telecom Cloud Market - Porter's Five Forces Analysis3.6.1. Supplier power3.6.2. Buyer power3.6.3. Substitution threat3.6.4. Threat from new entrant3.6.5. Competitive rivalry3.7. Telecom Cloud Market - PEST Analysis3.7.1. Political landscape3.7.2. Economic landscape3.7.3. Social landscape3.7.4. Technology landscape3.8. COVID-19 Impact Analysis

Chapter 4. Telecom Cloud Market Product Outlook4.1. Telecom Cloud Market, By Component Analysis & Market Share, 2021 & 20304.2. Solution4.2.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)4.2.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)4.3. Services4.3.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)4.3.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)4.3.3. Professional Services4.3.3.1. Market estimates and forecasts, 2017 - 20304.3.3.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)4.3.4. Managed Services4.3.4.1. Market estimates and forecasts, 2017 - 20304.3.4.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)

Chapter 5. Telecom Cloud Market Deployment Type Outlook5.1. Telecom Cloud Market, By Deployment Type Analysis & Market Share, 2021 & 20305.2. Private5.2.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)5.2.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)5.3. Public5.3.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)5.3.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)5.4. Hybrid5.4.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)5.4.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)

Chapter 6. Telecom Cloud Market Service Model Outlook6.1. Telecom Cloud Market, By Service Model Analysis & Market Share, 2021 & 20306.2. Software as a Service (SaaS)6.2.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)6.2.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)6.3. Platform as a Service (SaaS)6.3.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)6.3.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)6.4. Infrastructure as a Service (SaaS)6.4.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)6.4.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)

Chapter 7. Telecom Cloud Market Applications Outlook7.1. Telecom Cloud Market, By Applications Analysis & Market Share, 2021 & 20307.2. Network, Data Storage, and Computing7.2.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)7.2.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)7.3. Traffic Management7.3.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)7.3.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)7.4. Cloud Migration7.4.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)7.4.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)7.5. Others7.5.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)7.5.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)

Chapter 8. Telecom Cloud Market Enterprise Size Outlook8.1. Telecom Cloud Market, By Enterprise Size Analysis & Market Share, 2021 & 20308.2. Large Enterprises8.2.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)8.2.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)8.3. SMEs8.3.1. Market estimates and forecasts, 2017 - 2030 (USD Billion)8.3.2. Market estimates and forecasts, By Region, 2017 - 2030 (USD Billion)

Chapter 9. Telecom Cloud Market: Regional Estimates & Trend Analysis

Chapter 10. Competitive Analysis10.1 Recent Developments and Impact Analysis, by Key Market Participants10.2 Company/Competition Categorization (Key Innovators, Market Leaders, Emerging, Niche Players)10.3 Vendor Landscape10.3.1 Key company market share analysis, 202110.3.2 Company Analysis Tools10.3.3 Market Position Analysis10.3.4 Competitive Dashboard Analysis

Chapter 11. Competitive Landscape11.1. Company Profiles11.2. Juniper Networks, Inc.11.2.1. Company overview11.2.2. Financial performance11.2.3. Type benchmarking11.2.4. Strategic initiatives11.3. IBM Corporation11.3.1. Company overview11.3.2. Financial performance11.3.3. Type benchmarking11.3.4. Strategic initiatives11.4. Mavenir11.4.1. Company overview11.4.2. Financial performance11.4.3. Type benchmarking11.4.4. Strategic initiatives11.5. Affirmed Networks11.5.1. Company overview11.5.2. Financial performance11.5.3. Type benchmarking11.5.4. Strategic initiatives11.6. Fortinet11.6.1. Company overview11.6.2. Financial performance11.6.3. Type benchmarking11.6.4. Strategic initiatives11.7. Orange11.7.1. Company overview11.7.2. Financial performance11.7.3. Type benchmarking11.7.4. Strategic initiatives11.8. Huawei Technologies Co., Ltd.11.8.1. Company overview11.8.2. Financial performance11.8.3. Type benchmarking11.8.4. Strategic initiatives11.9. VMWare11.9.1. Company overview11.9.2. Financial performance11.9.3. Type benchmarking11.9.4. Strategic initiatives11.10. Cisco11.10.1. Company overview11.10.2. Financial performance11.10.3. Type benchmarking11.10.4. Strategic initiatives11.11. Nokia11.11.1. Company overview11.11.2. Financial performance11.11.3. Type benchmarking11.11.4. Strategic initiatives11.12. Ericsson11.12.1 Company overview11.12.2 Financial performance11.12.3 Type benchmarking11.12.4 Strategic initiatives

For more information about this report visit https://www.researchandmarkets.com/r/ikxoga

See the article here:
The Worldwide Telecom Cloud Industry is Expected to Reach $103.6 Billion by 2030 - Benzinga

Read More..

PIPEDA Findings #2022-004: Investigation into MGM breach highlights how to assess risk, and need for timely assessment – Office of the Privacy…

PIPEDA Findings #2022-004

19 May 2022

In February 2020, the Office of the Privacy Commissioner of Canada (the OPC) became aware of media reports regarding a large-scale data breach MGM suffered in 2019. Not having received a breach report on the matter, the OPC engaged with MGM to obtain additional information about the breach and the involvement of any personal information belonging to Canadians.

After receiving confirmation that Canadian personal information was affected by the breach, considering the potential impact on Canadians who were affected but had not yet been notified of the breach, and considering the significant passage of time since MGMs confirmation of the breach, the Commissioner initiated a complaint to investigateFootnote 1 whether MGM had complied with its mandatory breach reporting obligations under PIPEDA.Footnote 2

Our investigation found that MGM contravened the mandatory breach reporting provisions of PIPEDA. While MGM determined that there was a breach of its security safeguards in the summer of 2019, MGM failed to promptly assess whether the breach posed a real risk of significant harm (RROSH) to affected Canadians. We found that the breach posed a RROSH and that MGM did not report the breach to the Privacy Commissioner or notify affected individuals as soon as feasible.

In response to recommendations by our Office, MGM agreed that it would make amendments to its privacy breach response framework or process by 30 June 2022, to ensure that where MGM learns of a breach that may involve personal information of Canadian residents: (a) MGM will promptly conduct a RROSH assessment, consistent with the OPCs published guidance; and if MGM determines that such a breach gives rise to a RROSH, MGM will, as soon as feasible (b) provide a report to the Privacy Commissioner, and (c) notify affected individuals.

We therefore find the matter to be well-founded and conditionally resolved.

Under subsection 11(2) of PIPEDA, the Commissioner may initiate a complaint if he is satisfied that there are reasonable grounds to investigate a matter under Part 1 of PIPEDA.

Return to footnote 1

PIPEDA s 10.1

Return to footnote 2

ZDNet, Exclusive: Details of 10.6 million MGM hotel guests posted on a hacking forum, February 19, 2020; The New York Times, MGM Resorts Says Data Breach Exposed Some Guests Personal Information, February 19, 2020

Return to footnote 3

As discussed later in the report, the posting was removed from the hacker forum but later reappeared for sale on the dark web.

Return to footnote 4

Under subsection 11(2) of PIPEDA, the Commissioner may initiate a complaint if he is satisfied that there are reasonable grounds to investigate a matter under Part 1 of PIPEDA.

Return to footnote 5

PIPEDA s 10.1.

Return to footnote 6

Breach of Security Safeguards Regulations: SOR/2018-64.

Return to footnote 7

PIPEDA s 10.1(1).

Return to footnote 8

PIPEDA s 10.1(3).

Return to footnote 9

PIPEDA s 10.1(2) and s 10.1(6).

Return to footnote 10

An attempt by a third party to solicit confidential information from an individual, group, or organization by mimicking or spoofing, a specific usually well-known brand, usually for financial gain. Phishers attempt to trick users into disclosing personal data, such as credit card numbers, online banking credentials, and other sensitive information, which they may then use to commit fraudulent acts. See the Canadian Centre for Cyber Securitys glossary

Return to footnote 11

OPC, What you need to know about mandatory reporting of breaches of security safeguards (October 2018)

Return to footnote 12

Examples of OPCs published guidance: OPC, What you need to know about mandatory reporting of breaches of security safeguards (October 2018); and OPC, 2019 Breach record inspections (September 2020)

Return to footnote 13

Excerpt from:
PIPEDA Findings #2022-004: Investigation into MGM breach highlights how to assess risk, and need for timely assessment - Office of the Privacy...

Read More..

Hybrid cloud analytics: The future is now – ITWeb

Chris Pallikarides

Becoming a data-driven organisation in a world of exponential data growth could prove challenging and costly for most organisations. But hybrid cloud infrastructure offers the solution to future-proof analytics capability.

Moving to the cloud is no longer a question of if, but more about when, what and how? Having started with low-risk data analytical applications and low-hanging fruit, businesses are now moving more and more workloads to the cloud for agility and scale.

At the same time, companies in almost every industry are on a quest to become completely data-driven. Therefore, it is not enough to move only IT services and technology infrastructure to the cloud: to be truly data-driven, companies need to be able to analyse their data, wherever it may reside on-premises or across multi-cloud environments.

In a report, research company 451 Research highlights the following:

However, Gartner notes data and analytics activities now extend to the edge, across distributed devices, servers or gateways located outside data centres and public cloud infrastructure. Analysts estimate that by 2025, more than 50% of enterprise-critical data will be created and processed outside the data centre or cloud.

The intersection of hybrid cloud and data analytics should be a hybrid cloud analytics strategy underpinned by hybrid data warehousing or data lake technology, which will enable enterprises to support analysis of data where it resides on-premises or in the cloud.

Because they are so scalable, cloud environments are particularly appropriate for analytics involving massive volumes of data. They put less strain on on-premises systems, potentially at a lower cost.

Because they are so scalable, cloud environments are particularly appropriate for analytics involving massive volumes of data.

One client, a major online betting operator, processes between 200 million and 600 million records per day. Simply reporting on the days records took four or five hours a day using on-premises servers. Adding further analytics capabilities on-premises would put significant strain on systems and risk causing poor performance and downtime.

The organisation is now moving to a cloud environment, which will slash the time to compile reports, and the organisation will also be positioned to carry out more detailed analytics to support further innovation and growth.

Another client, a mid-sized manufacturer, has moved to a cloud environment to run analytics for better stock loss management. The companys on-premises infrastructure was not up to the demands of ongoing analytics, but a scalable cloud environment will allow it to gain insights to reduce losses and improve profits.

There are some important factors to consider before moving to a cloud-based or hybrid cloud analytics model: for one, data volumes are exploding, making it important to consider what data and how much of it will be hosted and processed in the cloud.

Organisations should determine what workloads existing systems manage, and whether analytics would put strain on them. If on-premises infrastructure is not up to the task, organisations should consider which workloads should be moved to which clouds, and how they will be integrated.

For high-speed data processing in the cloud, organisations will need to link multiple streams of data from different sources, such as the ERP, CRM and HR payroll systems, and big data platforms like Hadoop and Cloudera.

They should also consider stakeholder concerns around security and data sovereignty. Should these be a key priority, they should look to running on a private cloud, or in a public cloud with localised data centres, such as AWS or Azure.

Another factor to bear in mind is that poorly-considered cloud migrations can prove costly, and costs can run away with you if not properly controlled. It is important to set limits and governance rules to ensure cloud use doesnt scale over.

With a carefully-planned strategy, analytics in a hybrid cloud environment can deliver significant value and a faster ROI. Because the cloud is agile and scalable, environments can be spun up and down as needed, meaning organisations will always have access to sufficient capacity, with no wasted Capex and no risk to daily operations.

Go here to read the rest:
Hybrid cloud analytics: The future is now - ITWeb

Read More..

Accelebrate Joins AWS Training Partner Program to Deliver AWS Cloud Skills Training to IT Professionals online and at client sites – PR Web

Steve Heckler, Accelebrates president, commented We are honored to partner with AWS to teach AWS Cloud Practitioners, Architects, Developers, DevOps Engineers, and others how to take full advantage of the worlds most powerful cloud computing platform.

ATLANTA (PRWEB) September 28, 2022

Accelebrate Joins AWS Training Partner Program to Deliver AWS Cloud Skills Training to IT Professionals online and at client sites.

On August 30, 2022, Accelebrate, an IT training firm offering customized, private technical training, announced it has been approved by Amazon Web Services (AWS) to deliver online and on-site authorized AWS Training to learners at companies, government agencies, and other organizations. Accelebrate has joined the AWS Training Partner Program, which will enable Accelebrates AWS Authorized Instructors to deliver training developed by AWS.

As cloud technologies continue to help organizations transform their businesses at a rapid pace, employees with the necessary cloud skills are in high demand. Research shows that 85% of organizations report deficits in cloud expertise (451 Research, Voice of the Enterprise (VotE): Cloud, Hosting & Managed Services, Organizational Dynamics 2020, part of S&P Global Market Intelligence). The AWS Training Partner Program is designed for organizations like Accelebrate that meet or exceed rigorous criteria for delivering or offering high-quality technical training experiences. Accelebrate delivers AWS Training that enables IT professionals and businesses to leverage the power of the AWS Cloud for computing, data storage, and data science.

As an AWS Training Partner, Accelebrate delivers AWS training for all levels and roles, taught by expert instructors at your offices or online. Accelebrate prides itself on tailoring training to the needs of your learners and organization. This includes conducting an AWS Learning Needs Analysis (LNA) to assess the teams skills, identify gaps, and recommend training to close those gaps.

Accelebrates official AWS Classroom Training courses are taught by AWS Authorized Instructors with deep experience who help organizations to grow their AWS skill sets and prepare their staffs to pass the AWS certification exams. Accelebrate offers courses covering AWS Cloud Practitioner, Architect, Developer, Operations, Security, and Data Analysis and Machine Learning.

AWS Training is developed and maintained by AWS experts, ensuring the content reflects current best practices. AWS Classroom Training gives learners the opportunity to engage live and get questions answered by an expert instructor. Many courses also include hands-on labs, which allow learners to practice real-world scenarios in a sandbox environment. Training also helps prepare learners for AWS Certification exams, which validate technical skills and expertise with an industry-recognized credential.

Steve Heckler, Accelebrates president, commented We are honored to partner with AWS to teach AWS Cloud Practitioners, Architects, Developers, DevOps Engineers, and others how to take full advantage of the worlds most powerful cloud computing platform. We are especially looking forward to supporting our data science training clients as they shift their workloads to the cloud and take advantage of native features unique to the AWS platform.

Accelebrate also offers AWS Certification vouchers bundled with AWS Training courses to make it easier for customers looking to validate cloud skills.

Organizations need individuals with cloud skills to help transform their business, and there is a growing demand for IT professionals with AWS Cloud skills. AWS Training and Certification, along with our AWS Training Partners like Accelebrate, aims to equip the builders of today and tomorrow with the knowledge they need to leverage the power of the AWS Cloud. AWS Training, designed by the experts at AWS, teaches in-demand cloud skills and best practices, and helps learners prepare for AWS Certification exams so they can advance their careers and transform their organizations.

As cloud technologies continue to help organizations transform their businesses at a rapid pace, employees with the necessary cloud skills are in high demand. The AWS Training Partner Program is designed for organizations like Accelebrate that meet or exceed rigorous criteria for delivering or offering high-quality technical training experiences. As part of the program, Accelebrate delivers AWS Training to enable IT professionals and businesses to leverage the power of the AWS Cloud.

Maureen Lonergan, Director, AWS Training and Certification

About AccelebrateIn 2002, Steve Heckler founded Accelebrate with one laptop and himself as the sole instructor. In the following 20 years, Accelebrate has grown to consistently deliver more than 1,000 days/year of private training for over 15,000 employees, with classes delivered in-person and online for attendees from every US state, every Canadian province, and more than 30 countries. Media Contact:

Website: http://www.accelebrate.comEmail: info@accelebrate.comPhone from US and Canada: tel:8778491850Phone worldwide: tel:16786483113LinkedIn: https://www.linkedin.com/company/accelebrateFacebook: https://www.facebook.com/accelebrate/Twitter: https://twitter.com/accelebrateYouTube: https://www.youtube.com/c/accelebrateinc

Share article on social media or email:

Visit link:
Accelebrate Joins AWS Training Partner Program to Deliver AWS Cloud Skills Training to IT Professionals online and at client sites - PR Web

Read More..

The Global Cloud Native Storage Market size is expected to reach $41.9 billion by 2028, rising at a market growth of 22.3% CAGR during the forecast…

ReportLinker

A new approach to create and run software applications is termed, cloud native. Some examples of cloud-native include cloud computing, containerization, serverless architectures, and microservices.

New York, Sept. 29, 2022 (GLOBE NEWSWIRE) -- Reportlinker.com announces the release of the report "Global Cloud Native Storage Market Size, Share & Industry Trends Analysis Report By Component, By Deployment Type, By Organization Size, By Vertical, By Regional Outlook and Forecast, 2022 - 2028" - https://www.reportlinker.com/p06321920/?utm_source=GNW A storage solution created specifically for use in a cloud-native environment is known as cloud-native storage.

In cloud-native environments based on Kubernetes or other cloud native infrastructure, a cloud-native storage platform offers responses to ongoing data storage concerns as well as data management for domain-specific applications. New product storage technology, block storage, or conventional disc drives can all be used as the basis for object storage solutions in a distributed architecture.

The capacity to access data in the event of a breakdown, whether it be in the storage medium, controller, transmission system or any other component, is known as storage system availability. Storing redundant copies of data on another storage device, managing failover to redundant devices in case of failure, and fixing and restoring broken components are the three components of storage availability.

An increase in the investment in cutting-edge technology is the main element driving the growth of the market for cloud-native storage. Due to the necessity of Kubernetes, the cloud infrastructure stack has been greatly reduced in terms of cloud native storage and object storage, which allows network as well as portable storage.

COVID-19 Impact Analysis

Compared to the pre-COVID-19 situation, a greater demand for file storage systems is anticipated. During the forecast period, it is anticipated that the impact of Covid-19 will cause a noticeably high rate of market growth. As businesses transition to remote working environments, employees may collaborate & stay connected owing to cloud-based solutions. With the growing use of collaboration as well as conferencing tools by the distant workforce, the COVID-19 pandemic hastened the adoption of cloud-based storage.

Market Growth Factors

Rising Need Of Automation In Application Updates

Businesses can build automated platforms for constant testing and deployment using cloud native architecture. Every six months or so, banks would conduct system changes that were very carefully planned out, and they would tell consumers in advance of any projected downtime by publishing on banking websites. The majority of banking and financial apps, on the other hand, would probably update every week or even each day if they had native cloud storage, without having to halt operation or alert users.

Faster Data Backup And Recovery Growing The Demand For Cloud Native Storage

Automation and adaptability in cloud native storage boost flexibility, dependability, and availability. The advantage of automation in cloud native storage is that if something goes wrong, lost data can be quickly restored without having to halt the service. Data handling with block storage is made easier by using cloud native storage, which includes features such as cross-cluster disaster recovery, automated updating, as well as volume encryption.

Market Restraining Factors

Network Use And Latency Problems With Public Cloud

There are latency issues with the public cloud while storing & receiving data. Losing end-to-end network control when data is collected in the public cloud, together with high latency, can have a disruptive impact on how apps operate. Similar network troubles on the side of the provider of cloud native storage could interrupt storage solutions, which eventually causes work to stop and major losses for businesses. Additional sources of latency issues include insufficient Internet bandwidth, remote data center locations, and high network traffic.

Component Outlook

Based on component, the cloud native storage market segmented into solutions (file storage, block storage, object storage), and services. In 2021, the solutions segment dominated the cloud native storage market by generating the maximum revenue share. This is because they provide unstructured data storage without the scaling issues associated with conventional file storage. Instead of managing data as files or folders, object storage handles data as objects, which are stored in a single, sizable repository that may be dispersed across several physical storage devices.

Organization Size Outlook

On the basis of organization size, the cloud native storage market is fragmented into large enterprises, and SMEs. In 2021, the large enterprise segment held the highest revenue share in the cloud native storage market. In comparison to SMEs, major organizations have adopted cloud native storage more quickly. The expansion of new technology, goods, services, and solutions is the cause of the low adoption costs. This is leading to the demand for cloud-native storage among SMEs and big businesses.

Deployment Type Outlook

By deployment type, the cloud native storage market is divided into public cloud and private cloud. The private cloud accounted a substantial revenue share in the cloud native storage market in 2021. Private cloud storage is frequently provisioned utilizing out-of-date storage infrastructure. The private cloud should perform the same for storage.

Vertical Outlook

Based on vertical, the cloud native storage market is classified into BFSI, government, healthcare, telecommunication & IT, manufacturing, energy & utilities, media & entertainment, retail & consumer goods, and other verticals. The telecom & IT segment generated a significant revenue share in the cloud native storage market in 2021. Many operators, as well as vendors, are adopting cloud native technology in the telecommunications industry. Although they have long been used by public cloud and IT businesses, the telecoms sector nevertheless faces some unique difficulties.

Regional Outlook

Region wise, the cloud native storage market is analyzed across North America, Europe, Asia Pacific and LAMEA. In 2021, the North America region led the cloud native storage market by generating the largest revenue share. This is due to the rapid adoption of cloud native storage in the nations such as the United States & Canada. Additionally, the regions growth is anticipated due to the presence of top vendors in the region including Microsoft, AWS, IBM, and Google. Also, significant technological developments are taking place in the region supporting the market growth.

The major strategies followed by the market participants are Partnerships. Based on the Analysis presented in the Cardinal matrix; Microsoft Corporation and Google LLC are the forerunners in the Cloud Native Storage Market. Companies such as Amazon Web Services, Inc. (Amazon.com, Inc.), Alibaba Group Holding Limited, Huawei Technologies Co., Ltd. (Huawei Investment & Holding Co., Ltd.) are some of the key innovators in Cloud Native Storage Market.

The market research report covers the analysis of key stake holders of the market. Key companies profiled in the report include Microsoft Corporation, IBM Corporation, Amazon Web Services, Inc. (Amazon.com, Inc.), Google LLC, Alibaba Group Holding Limited, VMware, Inc., Huawei Technologies Co., Ltd. (Huawei Investment & Holding Co., Ltd.), Citrix Systems, Inc., Rackspace Technology, Inc., and Splunk, Inc.

Recent Strategies Deployed in Cloud Native Storage Market

Partnerships, Collaborations & Agreements:

Jun-2022: Splunk partnered with JupiterOne, the industrys leading cyber asset attack surface management (CAASM) platform. The partnership would allow users to gain context from a single source of truth across their entire security operations and conduct compliance assessments across their cyber assets and environment.

May-2022: Alibaba Group came into partnership with STC, the Saudi digital enabler of telecommunications services. This partnership would leverage the position of the STC Group in the services & technology solutions sector. It would employ the regions top talent & expertise to support the local capabilities of the kingdom. The enterprise would build an advanced ecosystem capable of meeting the future needs of the Middle East.

May-2022: VMware joined hands with Wipro, a leading global information technology, consulting, and business process services company. This collaboration would allow consumers to achieve the cloud freedom they desire with the enterprise control they require as they execute their digital strategies. Under the collaboration, the companies would bring together the power of VMware Cross-Cloud services with industry-leading Wipro FullStride Cloud Services to help global enterprises accelerate app modernization and reduce the cost, complexity, and risk of moving to the cloud.

Apr-2022: Google Cloud partnered with Linux Foundation, the nonprofit organization enabling mass innovation through open source. Under this partnership, the companies together would work on the project, Nephio. This would result in faster and better connected cloud-native networks of the future.

Mar-2022: Huawei came into partnership with ZainTech, a leading regional digital and ICT solutions provider. This partnership aimed at enhancing its digital & cloud services offerings across various industries in the region. With this partnership, the company would boost its technical innovation & guarantee a superior consumer experience for both its enterprise & government consumers in oil & gas, cloud, 5G campus, big data, smart public transport, and smart city solutions.

Nov-2021: Microsoft came into a partnership with Veritas Technologies, an American international data management company. This partnership aimed at expanding Veritas Enterprise Data Services with Microsoft Azure and simplifying data management while decreasing IT complexity for the enterprise by delivering solutions built on Microsoft Azure.

Nov-2021: AWS teamed up with Goldman Sachs, a leading global investment banking, securities, and investment management firm. Through this collaboration, the companies aimed at creating new data management and analytics solutions for financial services organizations. This collaboration would redefine how clients can discover, organize, & analyze data in the cloud, thus gaining rapid insights as well as driving informed investment decisions.

Nov-2021: VMware extended its existing partnership with Kyndryl, an American multinational information technology infrastructure services provider. This expansion aimed at accelerating IT & business reinvention for consumers through the combination of VMware solutions & Kyndryls design, build and managed services. The companies further aimed to help consumers speed their digital transformations by rapidly building & deploying new, more secure applications designed and built for a world of distributed work.

Oct-2021: VMware teamed up with Telia Company, the leading information technology, communications, and media services provider. This collaboration focused on accelerating Telias transition to a cloud-native infrastructure as an integral part of Telias company-wide transformation across the markets.

Jun-2021: Huawei partnered with Temenos, the banking software company. Under this partnership, the companies would offer Temenos cloud-native core banking solution on the Huawei Public Cloud. Furthermore, the partnership would bring together the extensive implementation, cloud hosting, and integration strengths of Huawei and the power of Temenos industry-leading banking software.

Sep-2020: Rackspace Technology collaborated with Dell Technologies, an American multinational technology company. This collaboration aimed at creating new modernized solutions for multi-cloud. The companies together would help solve some of the most pressing business transformation challenges facing organizations today, which include optimizing their data center footprints, controlling IT spending, removing barriers to multi-cloud integration, and dealing with security & compliance across a multi-cloud environment.

Product Launches and Product Expansions:

Jun-2022: Alibaba Cloud introduced Cloud Infrastructure Processing Unit (CIPU), a new cloud infrastructure system designed in-house to power its cloud-native data centers during its annual summit. This CIPU helps Alibaba Cloud deliver performance enhancements in networking, storage, security, and computing power to clients by offloading virtualization functions from servers to dedicated hardware.

May-2022: Microsoft extended the Azure cloud platform by adding a slew of updates. The major update is the availability of Draft 2, a reboot of the Draft open-source project that helps developers build Kubernetes apps. Draft 2 serves with an easy way to create, containerize & employ apps directly on Kubernetes, and can also generate GitHub Action workflow files.

Nov-2021: Microsoft extended its Azure Container Apps by introducing a new cloud-native offering. This expansion would enable developers to build microservice architectures using containers. Azure Container Apps is the only serverless container solution offering a combination of both built-in HTTP-based autoscaling, including scaled to zero, and event-driven-based autoscaling to run containers that may be processing messages from queues, streams, or databases.

Sep-2021: Amazon Web Services released Amazon FSx, a new technology for NetApps Ontap. Amazon FSx would give NetApp & non-NetApp consumers the chance to run the worlds best storage platform in AWS.

Sep-2021: Google launched Filestore Enterprise & Backup for Google Kubernetes Engine (GKE). The launch would make it easier for consumers to protect data out-of-the-box, across a wide variety of applications & use cases.

Apr-2021: IBM released IBM Spectrum Fusion Hyper-Converged Infrastructure, advanced storage solutions. The company aimed at simplifying data accessibility & availability across hybrid clouds along with planning to ship new container-native software-defined storage solutions designed to help companies expand data availability across complex hybrid clouds for greater business performance and resilience.

Acquisition and Mergers:

Feb-2022: IBM completed the acquisition of Sentaca, a telco consulting services provider. Through this acquisition, the company focused on bolstering its position in the cloud-native & hybrid-cloud market for wireless operators. This acquisition would strengthen IBMs ability to help carriers modernize their infrastructure & critical applications with its open source platforms like Red Hat OpenShift and OpenStack.

Jan-2022: Rackspace Technology acquired Just Analytics, a leading provider of cloud-based data, analytics, and Artificial Intelligence (AI) services. This acquisition would bring strong regional ties into the Microsoft Azure ecosystem with Just Analytics recently being awarded four regional Microsoft partner of the year awards.

Mar-2021: VMware completed the acquisition of Mesh7, cloud-native application security. The acquisition aimed at boosting the companys Kubernetes, micro-services & cloud-native capabilities. The acquisition of Mesh7s contextual API behavior security solution with Tanzu Service Mesh would enable VMware to deliver a high fidelity understanding of which application components are talking to which using APIs.

Geographical Expansions:

Jul-2022: Google expanded its geographical footprints in Latin America and Mexico. Through this expansion, the company focused on expanding its digital infrastructure, supporting digital skills, fostering an entrepreneurial ecosystem & helps create inclusive and sustainable communities. Further, the company aimed at fulfilling the rising demand for cloud services across the globe.

Oct-2021: Alibaba Cloud expanded its geographical footprints in South Korea and Thailand by setting up new data centers in the regions. The company focused on assisting local enterprises in their digital innovation journey, along with unveiling a slew of products and solutions. This acquisition aimed at addressing local consumers in South Korea & Thailand to meet their digitalization requirements from both a technology and services perspective.

Scope of the Study

Market Segments covered in the Report:

By Component

Solution

o Block Storage

o File Storage

o Object Storage

Services

By Deployment Type

Public Cloud

Private Cloud

By Organization Size

Large Enterprises

SMEs

By Vertical

BFSI

Telecom & IT

Healthcare

Retail & Consumer Goods

Manufacturing

Government

Energy & Utilities

Media & Entertainment

Others

By Geography

North America

o US

o Canada

o Mexico

o Rest of North America

Europe

o Germany

o UK

o France

o Russia

o Spain

o Italy

o Rest of Europe

Asia Pacific

o China

o Japan

o India

o South Korea

o Singapore

o Malaysia

View post:
The Global Cloud Native Storage Market size is expected to reach $41.9 billion by 2028, rising at a market growth of 22.3% CAGR during the forecast...

Read More..

Banks Stumble To Their Cloud-Based Future – Forbes

Banks' cloud migration efforts could use a lot of help.

In a January 2022 article titled Banks Tiptoe Toward Their Cloud-Based Future, the New York Times wrote:

The banking industry has been mostly slow to adopt cloud computing. While Wall Street leaders have long acknowledged the potential of cloud computing to cut costs, they have only allowed their firms to take halting steps. Some firms are held back by old computer systems that are difficult to revamp or retire, making the transition even more tricky.

Adoption isnt the issue and banks arent tiptoeing anywhere.

According to Cornerstone Advisors 2022 Whats Going On in Banking study, two-thirds of US banks and credit unions already haveor expect to have by the end of 2022apps running in the cloud.

The challenge is transitioning to the cloud and replacing systems and applications that could increase efficiency and effectivenessand maybe even provide competitive advantageby being in the cloud.

Nine in 10 financial institutions have a digital transformation initiative in place, and 80% of them believe theyre not even halfway done with transformation efforts. How rapidly they can migrate to the cloud determines how fastand effectivelythey can complete their digital transformation.

A team of Dutch university professors assessed the costs of cloud migration at 10 international corporate banks. The researchers identified nine cost categories associated with cloud migrations:

Cloud migration cost categories

In five of the categoriesdependencies, legislation, departmental support, re-architecting, and external contractorsat least half of the banks experienced cost overruns (versus original estimates) for their cloud migration efforts.

Cloud migration cost overruns

Application dependencies were the most common category of budget overruns. One banker commented:

Public cloud initially starts on-premise and all your connections and dependencies are on-premise and to decompose them for the purpose of migration is more complicated than expected.

Another said:

We had to decompose some applications due to dependencies. Cloud adoption in practice is much slower than expected due these kind of complexities.

Although the study didnt quantify the overrun of cloud migration costs, the qualitative conclusions clearly point to a system underestimation of cloud expenses, driven by a few cloud migration cost categories.

Interestingly, only three of the 10 banks included in the Dutch study said they had cost overruns for their internal core migration team. Of the other seven, however, three had budget overruns on external resources.

One bank exec commented:

The level of experience within the industry is very low regarding cloud computing, the chance of hitting unexpected costs is high, because we are experiencing it for the first time.

Another added:

When we started we thought we would need a certain amount of resources. However, we miscalculated because we were too positive. We have a knowledge gap and cloud is the main reason for this.

And the researchers themselves remarked:

Overspending on external contractors is an indication how the process of migration is actually progressing. It displays which areas lack the required knowledge to manage the cloud migration internally.

One banker speculated the banks staff would leave if cloud migration was too slow:

Our bank worked for many years with external contractors, they all left and the current internal employees are recently hired. We have very limited knowledge of the on-premise IT infrastructure and this is a continuous struggle for us.

According to Cornerstones research, nine in 10 financial institutions have a digital transformation initiative underway, and 80% of them believe theyre not even halfway done with their transformation efforts.

Theres a connection here between cloud migration and digital transformation progress. The slower the migration to the cloud, the slower the digital transformation.

Like it or not, for many banks, the key to accelerating digital transformationand cloud migrationis more effective vendor participation.

When discussing partnerships, bank executives cite a lack of coordination and transparency of third party vendors as a concern.

According to another report from Cornerstone Advisors, commissioned by Amdocs, titled Leveraging the Cloud to Accelerate Digital Transformation:

As banks integrate more and more systems into their core in order to modernize, it means more partners to manage. Having a dedicated vendor manager is mission critical to manage all of the relationships as well as designating someone within the financial institution to sit in the captains chair to drive the initiative.

In addition to better management of external resources, Cornerstone advises financial institutions to transform the IT talent stack:

The talent shortage will continue to slow down cloud implementation. This forces banks to look for talent more aggressively in key areas including specific cloud platform configuration (AWS, Azure, Google), server virtualization and app containerization, API integration, DevOps, cloud security and orchestration, and disciplined change management.

Banks may be struggling with their cloud migration efforts, but thats preferable to where the industry was just a few years ago. In a 2019 report titled The Inevitable Journey to the Cloud in Banking, I wrote:

[Cloud] proponents tout benefits like increased reliability, reduced costs and the transformation of fixed costs into variable costs (capital expense vs. operational expense), while opponents cite concerns regarding security, data privacy and migration challenges.

The proponents may have been too optimistic about cost reduction and transformation and the opponents overly worried about the security implications. The two sides are converging, however, and banks are moving to the cloud. Even if they are stumbling their way there.

For a complimentary copy of the Cornerstone Advisors report, Leveraging the Cloud to Accelerate Digital Transformation, click here.

Read more:
Banks Stumble To Their Cloud-Based Future - Forbes

Read More..

How cloud computing will serve as the foundation for business innovation – Gulf Business

For the last few years, cloud adoption has seen an upward trajectory in the Middle East, with many organisations moving data to the cloud. The region is undergoing an economic transformation, with many nations implementing smart city initiatives and updating their national visions while improving their digital economies. National visions in countries such as UAE, Saudi Arabia, Bahrain and Kuwait encourage governments to use emerging technologies in order to achieve their transformation objectives.

For instance, the digital transformation of government operations is a key component of Saudi Arabias Vision 2030. A cloud-first strategy aids lowering total cost of ownership, and strengthening cybersecurity.

Meanwhile, Bahrains cloud-first policy is also devoted to modernising government ICT and setting the bar high for employing cloud computing services in order to cut costs, boost security and efficiency and create first-rate citizen services.

Accelerating cloud adoptionPublic and private sectors in the region are investing in cloud computing to improve their infrastructure and services and leverage this technology to organise and streamline their processes.

Going by the numbers, Gartner forecasts end-user spending on public cloud services in the Middle East and North Africa (MENA) to grow 19 per cent this year. A renewed focus on technology growth post-Covid-19 in the region is leading to continued growth in public cloud spending, said Colleen Graham, senior research director at Gartner, in a statement. Various MENA governments policies on telemedicine, usage of autonomous vehicles, smart cities, and a rapid move towards the next phase of the fourth industrial revolution are opening new growth avenues for public cloud in the region. Additionally, the attention given to building and nurturing talent will turn a new leaf in the regions shift towards becoming a digital economy.

Gartner also stated that in 2022, MENA chief information officers will spend the most on cloud application services, which include business intelligence applications, email and authoring, content services, customer experience and relationship management and supply chain. This segment will total $2.3bn, an increase of 16 per cent from 2021, and will account for 40 per cent of the total investment made in public cloud services.

The second largest segment will be cloud application and infrastructure services, which is forecast to total $1.1bn in 2022, an increase of 25.8 per cent from 2021. Cloud system infrastructure services will record the highest growth. This segment will grow 36.8 per cent to a total $895m in 2022. Dario Sarmiento, manager Service Delivery at Kyndryl Gulf and Levant, believes that disruptive technologies have been at the centre of the IT industry for decades. However, their adoption has recently increased at an accelerated pace, forcing businesses to work in and keep up with a very competitive environment.

In the last five years, we have seen how the convergence of cheaper network access, coupling with the development of better mobile and wireless technologies (peaking, for example, at the current promise of 5G speed for wireless communications) and improvements in the underlying virtualisation technologies, has created multiple options for enterprises to enhance their competitiveness. On the flip side, with all thats available today, some businesses are finding it challenging to prioritise IT initiatives based on their own strategic objectives. Cloud Computing is part of any of those relevant options, so a decision to move there cannot be disregarded, notes Sarmiento.

Are some companies reluctant to move to the cloud?As the world becomes technology-driven, cloud computing has become an integral part of daily life, especially for regional businesses. The cloud helps companies organise their data and keep a backup of every single piece of information used for product development and customer experience enhancement. However, many organisations are still putting off migrating their infrastructure to the cloud. Reason while some may lack IT personnel, others fear the loss of accessibility. Furthermore, the decision-makers focus on the cost. Meanwhile Fayez Eweidat, senior director META at Juniper Networks, observes security as one of the biggest concerns surrounding public cloud adoption. As more companies empower a work-from-anywhere workforce, access to cloud-based applications provides a seamless experience from home to office. However, some companies are hesitant to walk down this path due to concerns about how to transition their existing security, policies and frameworks and continue to meet their compliance commitments, he adds.

However, Sarmiento says it could be one or a combination of various reasons. We have seen organisations wanting to move to the cloud but are held back by the rigidity of their legacy systems. The reality is that the larger the organisations, the longer the legacy systems have been in place. Another stumbling block is the shortage of cloud skills, which can easily impede cloud transformation projects.

The risks of not moving to the cloudSo, what risks are associated with an organisation not moving to the cloud? With the cloud at the centre of all digital transformation journeys, businesses may not be able to capture more extensive opportunities, and they could face limitations on scalability, where the cost would be too high for them to grow because of traditional IT spending that utilises funds that the business could instead use for growth initiatives or innovation.

Not being part of the cloud community also reduces the intelligence of a business. By that, I mean that in todays corporate world, startups and SMEs need to have a smart approach to data. Business intelligence and data analytics need a good understanding of data collection and storage. Not having a robust cloud storage system supporting your business can, therefore, leave a negative impact on having to dig into your data, making your business less efficient, explains Candid Wuest, vice president Cyber Protection and Research at Acronis.

However, before investing in the cloud, organisations should consider how it aligns with their overall business strategy and understand which workloads are best suited to run on the cloud. In short, investing in the cloud must be aligned with developing strategic initiatives. Primarily, an organisation should consider security, compliance, hidden costs, networking, complexities and performance before investing in a cloud environment. Additionally, the preparedness of the team to manage not only the technical aspects of migration and ongoing operations but also the business considerations of optimising costs and assuring compliance are critical factors. Lastly, organisations must assess the technical readiness of any application before moving it to the cloud, comments Eweidat.

Cloud security Tools and best practices Although the cloud is constantly evolving, some best practices have remained firm for ensuring the security of cloud environments. Experts suggest that organisations with existing cloud solutions in place or looking to implement them should consider the advice below to ensure data security.

Emad Haffar, head of Technical Experts at Kaspersky, explains that basic antivirus and anti-malware protection are not enough to protect a cloud infrastructure. Industry best practices dictate that every operating system in the infrastructure needs comprehensive, multi-layered protection that safeguards various types of workloads running on different platforms. Understanding the difference between the cybersecurity responsibility of a cloud host and a cloud tenant is vital as well. In a private cloud setup, protecting the underlying virtualisation and storage as well as network connectivity, must be considered. And while this responsibility is on the shoulders of the provider in a public cloud setup, workload security is still needed, and it is on the tenant side. He adds that having a comprehensive security solution that can cater to all these different scenarios can make the task more manageable.

Meanwhile, Eweidat puts focus on zero trust on how security teams should redesign networks into secure micro perimeters, strengthen data security using obfuscation techniques, limit the risks associated with excessive user privileges and access, and dramatically improve security detection and response with analytics and automation. Organisations investing in security tools and architectures should ensure their investment focuses on managing and monitoring of cloud environments, actionable insights for faster threat detection, security behaviour and event-driven capabilities for real-time alerts.

OverviewMoving forward, the cloud will drive technological innovation and serve as the foundation for business innovation. Thanks to the rising user demand, evolving needs of the organisations and the massive quantity of data, the future of cloud computing in the Middle East is on a steady path to proliferate in the coming years.

Read: Case study: How UAEs Mercury Payments Services fast-tracked its cloud migration with Oracle

The rest is here:
How cloud computing will serve as the foundation for business innovation - Gulf Business

Read More..