Page 4,097«..1020..4,0964,0974,0984,099..4,1104,120..»

Has Arm Discovered the Ecosystem Keys? – The Next Platform

Arm server development is a reality and a growing one at that. Not just from a performance point of view but also, perhaps more important, from an ecosystem view.

Be it the Marvell ThunderX2 processor or the Ampere eMAG Skylark processor, the hyperscale, cloud, enterprise ecosystems are willing to adopt these new processors to further improve their TCO or dollars/core.

The all-important ecosystem is catching up with Arm, which is key to the momentum necessary to make the Arm servers a sustainable reality. With AWS launching their version of Arm instances i.e. Graviton processors, theres the much needed push to make the software ecosystem more widely acceptable in the industry. Not just that, AWS even announced bare-metal offerinings for EC2 A1 instances.

Slowly but steadily, Arm has also made a mark for itself in high performance computing, something we expect to see in full force at this years Supercomputing Conference. Arm has the most traction in terms of deployments and software development in HPC in the United States, Europe and Japan with each region leading the way along different trajectories to deploy systems based on the Arm architecture for their supercomputers.

All of this has taken time and extended development, of course. The first wave of Arm based servers came in 2010 until 2014 and were more experimental in nature than real production systems.

The first 64-bit Arm design i.e. the ARMv8-A was introduced in 2011 and since then the Arm server ecosystem have seen lots of ups and downs. ZTSystems, in November 2010 had launched a 1U Data Center Arm server based on Cortex-A9 cores (32-bit) which was supposed to be energy efficient and a denser solution compared to Intel Servers. Then came Calxeda with their version of 32-bit Arm servers i.e. the EnergyCore-ECX-1000 which did not see adoption and Calxeda eventually went defunct in 2013. In 2011 AppliedMicro launched the X-Gene 1 processor followed by X-Gene 2 in 2014. Samsung, Cavium (now Marvell) and AMD came up with their versions of Arm processors which tried to penetrate the server market but could not generate tangible interest among the end-users to adopt these technologies.

Arm servers have undergone a transformation in terms of development and early signs of this were seen in a semi-secret project within Broadcom which was taking shape in the form of Project Vulcan. The idea was to develop a world class 64-bit serious Arm server to take on Intel in the HPC and cloud market.

In late 2016, when Avago gave up on Broadcoms ambitions to develop a first class Arm server, Cavium jumped in and brought the Vulcan IP and Team on-board and fully funded the Vulcan project, re-christened as Cavium ThunderX2 now, Marvell ThunderX2. In more ways than one, the ThunderX2 is a serious contender to Intel and AMD in the HPC, hyperscale and cloud businesses.

To make things better for the Arm ecosystem, in 2017, a brand new company, Ampere Computing bought the X-Gene assets and re-introduced the X-Gene processor as the Ampere eMAG processor. It needs to be mentioned that Qualcomm tried its hand at building a true Data Center Arm Server Centriq based on the Falkor Architecture and given Qualcomms standing, with time, it could have made their data center server project a success. However, for reasons unknown to many, they chose to significantly disinvest and many personnel from Qualcomms Centriq project were hired by Ampere Computing in Raleigh. Huawei has a very compelling Arm Server offering in the Kunpeng 920, which is a 7-nm, 64 core CPU.

Figure 1: Diverse Arm architectures (source)

The question many have is whether the Arm server ecosystem is mature enough to be excited about?

The ecosystem has come a long way to become a stable one. However, it has many miles to go to reach the same level as x86. Given this momentum, it would not be surprising if the likes of Google, Facebook, Tencent etc. are actively experimenting with Arm platforms. Amazon and Microsoft have already invested in Arm platforms in their respective clouds i.e. AWS & Azure.

Figure 2: Commits to Linux GitHub repository for x86 vs. arm64 as of 13th November, 2019

The contributions towards enabling aarch64 for Linux operating system have steadily increased since 2012 while the growth rate for x86 has not been as consistent. These are good indications that the Arm ecosystem is here to stay and growing.

An ongoing debate among software engineers is whether to implement a business logic in a monolithic architecture or take the same logic and break it down into multiple pieces. There is a growing trend of organizations moving to a Microservices architecture for various reasons be it unit testing, ease of deployment, server performance among many others. Also, microservices based architecture are relatively easy to scale compared to a monolith. Linaro, Arm and Arm Server Manufacturers are leading this charge. Also, Packet is providing the developer community a platform to develop and sustain the ecosystem.

If theres one area where Arm servers have taken the biggest strides, it is definitely be High Performance Computing (HPC). The Arm ecosystem for HPC is also the most developed compared to Arms progress in cloud datacenters.

The momentum for Arm in HPC was driven by many centers, but Dr. Simon McIntosh-Smith and the University of Bristol and Cray hosting the 1st Isambard Hackathon to optimize HPC applications for ThunderX2 based servers back in November 2017 at Bristol. This was promptly followed up by a 2nd Isambard Hackathon in March 2018.

Most of the HPC applications compile and run out of the box for Arm based servers with Arm compilers, GCC, OpenMPI, OpenMP support.

I participated in both representing Cavium Inc, assisting developers, architects and engineers optimize their codes/applications for ThunderX2 Processors. Collectively, we optimized key HPC applications like NAMD, UM-NEMO, OpenFOAM, NWCHEM, CASTEP, etc. and compared to Intel CPU Architectures like Broadwell and Skylake. Prof Smith and team did a detailed study identifying the opportunities and benefits of Arm Servers with regards to the incumbent Intel servers with compelling performance per dollar for the Arm-based servers.

Figure 3: Cray-Isambard performance comparison on mini-apps

Figure 4: Cray-Isambard performance comparison on key Archer applications

Figure 5: Cavium Inc. published HPC Performance comparison vs. Intel Skylake CPUs (2017)

This was a significant movement that Arm servers needed in the HPC space. The two Isambard hackathons also fast-tracked the Arm HPC development with Arm optimizing their compilers as well as Math libraries in collaboration with Arm server manufacturers like Cavium Inc (now Marvell Semiconductors). There is tremendous movement in the Arm HPC Performance Libraries optimization world. Arm has invested in optimizing GEMM, SVE, spMM, spMV and FFT libraries in collaboration with developers and Silicon manufacturers like Marvell. The Arm Allinea Studio has successfully established itself as a go-to tool for Arm server Workload Analysis, similar to what VTune would be for Intel.

Another major milestone was the Vanguard Astra Arm based supercomputer at Sandia National Laboratories powered by DoE, Cavium and HPE. This is the first Arm based supercomputer to make the TOP500 list at 156th position as of June 2019 and 198th rank in the November 2019 rankings. The building blocks are HPE Apollo 70 platforms, Marvell ThunderX2 CPUs with 4xEDR Infiniband interconnect. The Astra Supercomputer is made up of 2592 compute servers i.e. 145k cores and 663 TB memory. US DoE is making a concerted effort to invest in diverse as well as future proof technologies such as Arm, in its path towards achieving exascale computing.

Figure 6: Astra, the Arm based supercomputer debuted on the TOP500 list in November 2018

Europe and Asia are taking huge strides in deploying Arm based clusters and systems for HPC and Research. Be it Monte-Carlo, Isambard or CINECA-E4 projects in Europe or Japans Arm based Fugaku supercomputer, its just the beginning of a new era of Arm in HPC. Cray is betting big with the A64FX Arm chip built by Fujitsu. The A64FX prototype is number one on the Green500 list and 160th on the Top500 list..

HPC workloads tend to be highly parallelizable in nature, and Arm CPUs provide an opportunity to leverage lots of cores at reasonable price points. Further, having competition in the CPU market benefits all buyers, not just HPC shops, to negotiate the best resources for their workloads.

Marvell is a pioneer in more ways than one in introducing the Arm server ecosystem to the hyperscale world with Marvell and Microsoft partnering on ThunderX2 platforms for Azure. Oracle has invested $40 Million in Ampere Computing, which is home to the ARMv8 eMAG processor. Oracle also has plans to massively expand their datacenter footprint in the coming months and this investment in Ampere could mean potential deployment of eMAG processors in Oracle Data Centers.

In the recent past, theres been a slew of announcements regarding enhancements to the Arm ecosystem. VMware announced 64-bit support Arm Support. In an official announcement, DDN announced professional support for Lustre on Arm servers in 2018 In mid 2019 at ISC, AMI announced firmware support for the Marvell ThunderX2 Arm based servers in March 2019.

NVIDIA announced CUDA support for Arm at ISC19 and backed it up with a major announcement of introducing a reference design to enable organizations to build GPU-accelerated Arm based servers, which is a big shift towards enabling Arm to be successful in the HPC and accelerated computing segment. Imagine a system with power efficient Arm based CPUs with GPUs for training and AI ASICs for inference. Machine Learning & Artificial Intelligence pose interesting opportunities & the collaboration with NVIDIA will enable this segment for Arm based solutions.

Like Intel, AMD and Arm, Ampere Computing too has created a developer program for developers to build and expand their Cloud Ecosystem. This will enable further and faster integration of Arm servers in the hyperscale and datacenter world in a much more open and collaborative way.

While the ecosystem still needs more time to grow and mature, it is steadily moving towards that nirvana of It just works. With the emergence of Arm in the computer architecture world along with RISC-V and many other semiconductor start-ups, its only a matter of time until aarch64 is the new normal like x86. That is what the community is all striving towards.

Once the developers are convinced that their software stack just works on Arm Servers, it would be a big win for the Arm Server ecosystem, and I for one am willing to make the bold claim that for many workloads especially HPC It just works

About the Author

Indraneil Gokhale is a Performance Engineer and leads the Hardware Engineering team at Box Inc. Indraneil has previously worked at Cavium (now Marvell), Uber and Intel. Indraneil has experience in optimizing HPC applications and workloads for x86 and aarch64 architectures. He has published white papers, book chapters on optimizing the Weather Research and Forecasting (WRF) application. Indraneil holds a Masters Degree in Electrical Engineering from Auburn University, USA and a Bachelors Degree in EEE from Jawaharlal Nehru Technological University, Hyderabad, India.

See the original post here:
Has Arm Discovered the Ecosystem Keys? - The Next Platform

Read More..

SME disaster recovery made easy with cloud, hybrid and HCI – ComputerWeekly.com

Disaster recovery(DR) must be reliable, speedy and economical. These are the basic requirements for all businesses, and small to medium-sized enterprises (SMEs) are no exception. But, for smaller firms, cost considerations will be at or near the top of the IT managers list.

Organisations often view disaster recovery as little more than insurance, or as IDC analyst Phil Goodwin puts it, an expense that is likely to have little payback.

Large firms in highly regulated markets such as financial services invest heavily in disaster recovery and business continuity, not least because they are mandated to do so.

RBS, for example, was fined 56m for an IT failure in 2012.

Smaller firms might not have the budget for DR, however. Or they might choose not to pay for it, hoping they can ride out any incident with backups and hard work. This is short-sighted.

The Uptime Institutes 2018 Global Data Center Surveyfound that 31% of businesses experienced downtime that caused severe damage, but that 80% of incidents could have been prevented.

Operating disaster recovery on a pay-as-you-go model, combined with faster data transfers over the public internet, is transforming DR options for SMEs

Disaster recovery is becoming cheaper, simpler and more effective, through the growth of cloud-based services.

Firms no longer need to invest in dedicated or duplicate hardware, remote datacentres and the skilled staff to maintain them. Cloud or hybrid cloud technologies allow smaller companies to outsource much of the technical side of DR provision and to move to an on-demand model.

Operating disaster recovery on a pay-as-you-go model, combined with faster data transfers over the public internet, is transforming DR options for SMEs.

For smaller firms, conventional backup and disaster recovery has long meant saving data to tape or other removable media and storing it off-site. In case of a disaster the firm would need to source new hardware and restore data and applications.

Larger or better-resourced businesses, or those depending heavily on their data, will have invested in automatic off-site data replication and even standby servers. Others contracted with specialist suppliers to manage DR.

In the past, DR meant having a redundant location that was either always-on, a hot DR facility, or the use of shared resources that would be configured and set up when needed, says Roy Illsley, a distinguished analyst at Ovum who covers infrastructure solutions. These solutions are typically expensive or bring long recovery times.

Some businesses have moved to ad-hoc cloud backup solutions, including consumer-grade online storage. But this is still labour-intensive.

Small and medium-sized enterprises often deem disaster recovery too expensive or peripheral to core operations to fully invest in, warns Mark Wass, a director at business continuity supplier Sungard.

They often settle for a DIY, cloud-based platform approach, which they believe to be cheaper, and then assign responsibility for managing it to the office manager, he says.

Although this avoids reliance on on-site backups, it is still a manual process. And, Wass says, assuming staff can rely on cloud-based storage and access to a personal device to work on is risky.

Fortunately, the market for online disaster recovery is growing, and so is the choice of suppliers. Companies such as Veeam, Rubrik, Commvault, Cohesity and Nutanix provide options for businesses that need an off-the-shelf solution.

Cloud-based DR is available either directly from suppliers, or through managed service providers and IT integrators.

OGL Computer is a managed service provider with 1,200 UK SME customers. It offers a choice of cloud recovery and cloud-based data replication, as well as dedicated recovery options for VMware and Hyper-V. Recovery allows firms to restore their key applications within 24 hours cloud-based replication provides recovery within seconds.

But, as OGLs enterprise solutions architect, Steve Bennett, says, there are also customers who need, or prefer, an on-premise solution.

The availability of cloud services is an obvious way backup is becoming easier for SMEs, says Bryan Betts, analyst at Freeform Dynamics. But if they cant or dont want to use cloud, theres also the appliance option, either physical or virtual.

The availability of cloud services is an obvious way backup is becoming easier for SMEs. But if they cant or dont want to use cloud, theres also the appliance option, either physical or virtual Bryan Betts, Freeform Dynamics

Add in the availability of flash storage and modern software that provides user self-service, and you can get a box that not only takes care of all or almost all your backups, but also lets users themselves quickly sort out most of their data recovery needs.

Changes in the way businesses buy technology are also affecting disaster recovery.

Hyper-converged infrastructure (HCI) started out as a way to make it easier to deploy virtual machines. But because hyper-converged systems usually include their own storage, they lend themselves to disaster recovery too. On-premise HCI systems can be out of the reach of SMEs, but the implementation of cloud-based HCI makes it much more accessible.

An example is Cohesitys Clusters, but Nutanix and Rubrik use similar technologies. Businesses with the in-house expertise can also look at replication to enterprise public cloud providers, such as Amazon Web Services, Microsoft Azure and Google Cloud Platform.

Cloud DR isnt [as] neatly packaged but is more flexible, says Peter Groucutt, managing director of backup provider Databarracks. You can choose your replication software, such as ASR, Zerto or Veeam, and choose your cloud platform. For day-to-day replication you can keep cloud resources to a minimum but have the resources available to scale up as much as you need when you invoke DR.

Another factor in favour of a bespoke approach is the growing need to back up cloud-based applications. The appliance-to-offsite-backup or appliance-to-cloud route is tried and tested, but data backups from cloud-based applications are largely best kept in the cloud.

IT managers should check their disaster recovery plan includes cloud applications. More software-as-a-service (SaaS) providers are including application availability and data protection options with their products.

No amount of technology, though, will protect a business if it fails to work. This means having a disaster recovery plan, a robust testing regimeand a plan to deal with human factors, from the availability of technical experts to senior managements ability to act under pressure.

You must have a robust process that has clear rules on when a switch to disaster recovery is required and when it is not Roy Illsley, Ovum

For SMEs, the main point to consider is ease of use in setting up the replication, and then how the DR capability can be tested and verified it is operating and fit for purpose, says Ovums Illsley. The area where most SMEs fail is in the process for invoking the DR plan. You must have a robust process that has clear rules on when a switch to DR is required and when it is not.

Freeform Dynamics Betts goes further. The absolutely key requirement for DR is to make sure you can recover, he warns. Test it often enough to make sure not just that your backup process is reliable, but that you can rebuild a working system from it within the specified recovery time objective and recovery point objective.

Read more:
SME disaster recovery made easy with cloud, hybrid and HCI - ComputerWeekly.com

Read More..

AllSaints tech team explains tools behind its cloud journey – Essential Retail

Members of the AllSaints technology team have explained the companys five-year journey to becoming a cloud-based business, describing it as the biggest infrastructure change in its history.

Since 2014, AllSaints has used the capabilities of data management specialist Pythian, API codifier Terraform, and an array of Google Cloud tools to help create a new infrastructure. During the project, the retailer has halved the number of servers it runs from 60 to 30, and it says it is in a stronger position to flex its online business to meet consumers ever-changing demands.

As previously reported by Essential Retail, AllSaints uses Google G Suite for internal communications, which the retailers executive consultant for digital and technology, John Bovill, said was a deliberate move to get staff familiar with using the cloud.

Soon after, AllSaints migrated its infrastructure to Google Cloud with support from Pythian. Andy Dean, technical operations manager at AllSaints, said 60 individual services moved onto the new technology set-up.

The interdependencies between them meant that it made more sense to move them all at once, and that took a lot of planning, he noted, adding that the development team needed to change to Cloud SQL technology and undertake significant re-coding.

It was the biggest infrastructure change wed made in the history of the company, so one of our goals was that nobody noticed the change, explained Dean.

One result of that change is AllSaints now utilises Google Compute Engine and Google Kubernetes Engine autoscaling to meet the needs of additional online traffic at peak times. It said this means it no longer has to rely on additional servers and has contributed to a 50% reduction in infrastructure costs.

The team now monitors web performance through Google Clouds Stackdriver, while Googles wider network provides disaster recovery services which replaces a previous reliance on a single data centre.

In-house development

AllSaints develops its customer-facing services in-house, including electronic point-of-sale systems and its mobile app, and the move towards a microservice architecture gave the development team a chance to change their way of working. They can now build a continuous integration/continuous deployment (CI/CD) pipeline to automate the software delivery process, using Jenkins on Google Cloud and Terraform.

Before, we couldnt confidently say a bug was fixed until we actually tested it in production now we can deploy code in test environments that exactly mimic production, noted Dean.

The improved CI/CD pipeline means we can update our services every day, with a shorter lifespan on bugs, and minimal disruption. That makes us more responsive to customer needs, more proactive. And that's exactly what were trying to achieve.

AllSaints said online page speeds and conversions have increased since rolling out the new infrastructure. Next up, the retailer is preparing to deploy Istio to connect and monitor its microservices model, and it is planning to explore ways of leveragingdata within its organisation via Google Cloud tool, BigQuery.

Strategically we are looking to maximise our usage of Google Cloud, driving this and associated technologies to provide the best possible AllSaints experience for our customers, explains Bovill, who having overseen the latter stages of the AllSaints cloud journey is preparing to leave the retailer at the end of this year.

See the rest here:
AllSaints tech team explains tools behind its cloud journey - Essential Retail

Read More..

Kubernetes – taming the cloud – TechRadar

When you want to use Linux to provide services to a business, those services will need to be secure, resilient and scalable. Nice words, but what do we mean by them?

Secure means that users can access to the data they require, be that read-only access or write access. At the same time, no data is exposed to any party thats not authorised to see it. Security is deceptive: you can think you have everything protected only to find out later that there are holes. Designing in security from the start of a project is far easier than trying to retrofit it later.

Resilient means your services tolerate failures within the infrastructure. A failure might be a server disk controller that can no longer access any disks, rendering the data unreachable. Or the failure might be a network switch that no longer enables two or more systems to communicate. In this context, a single point of failure or SPOF is a failure that adversely affects service availability. A resilient infrastructure is one with no SPOFs.

Scalable describes the ability of systems to handle spikes of demand gracefully. It also dictates how easily changes may be made to systems. For example, adding a new user, increasing the storage capacity or moving an infrastructure from Amazon Web Services to Google Cloud or even moving it in-house.

As soon as your infrastructure expands beyond one server, there are lots of options for increasing the security, resilience and scalability. Well look at how these problems have been solved traditionally, and what new technology is available that changes the face of big application computing.

(Image credit: Future)

Enjoying what you're reading? Want more Linux and open source? We can deliver, literally! Subscribe to Linux Format today at a bargain price. You can get print issues, digital editions or why not both? We deliver to your door worldwide for a simple yearly fee. So make your life better and easier, subscribe now!

To understand whats possible today, its helpful to look at how technology projects have been traditionally implemented. Back in the olden days that is, more than 10 years ago businesses would buy or lease hardware to run all the components of their applications. Even relatively simple applications, such as a WordPress website, have multiple components. In the case of WordPress, a MySQL database is needed along with a web server, such as Apache, and a way of handling PHP code. So, theyd build a server, set up Apache, PHP and MySQL, install WordPress and off theyd go.

By and large, that worked. It worked well enough that there are still a huge number of servers configured in exactly that way today. But it wasnt perfect, and two of the bigger problems were resilience and scalability.

Lack of resilience meant that any significant issue on the server would result in a loss of service. Clearly a catastrophic failure would mean no website, but there was also no room to carry out scheduled maintenance without impacting the website. Even installing and activating a routine security update for Apache would necessitate a few seconds outage for the website.

The resilience problem was largely solved by building high availability clusters. The principle was to have two servers running the website, configured such that the failure of either one didnt result in the website being down. The service being provided was resilient even if the individual servers were not.

Part of the power of Kubernetes is the abstraction it offers. From a developers perspective, they develop the application to run in a Docker container. Docker doesnt care whether its running on Windows, Linux or some other operating system. That same Docker container can be taken from the developers MacBook and run under Kubernetes without any modification.

The Kubernetes installation itself can be a single machine. Of course, a lot of the benefits of Kubernetes wont be available: there will be no auto-scaling; theres an obvious single point of failure, and so on. As a proof of concept in a test environment, though, it works.

Once youre ready for production, you can run in-house or on a Cloud provider such as AWS or Google Cloud. The Cloud providers have some built-in services that assist in running Kubernetes, but none of are hard requirements. If you want to move between Google, Amazon and your own infrastructure, you set up Kubernetes and move across. None of your applications have to change in any way.

And where is Linux? Kubernetes runs on Linux, but the operating system is invisible to the applications. This is a significant step in the maturity and usability of IT infrastructures.

The scalability problem is a bit trickier. Lets say your WordPress site gets 1,000 visitors a month. One day, your business is mentioned on Radio 4 or breakfast TV. Suddenly, you get more than a months worth of visitors in 20 minutes. Weve all heard stories of websites crashing, and thats typically why: a lack of scalability.

The two servers that helped with resilience could manage a higher workload than one server alone could, but thats still limited. Youd be paying for two servers 100 per cent of the time and most of the time both were working perfectly. Its likely that one alone could run your site. Then John Humphrys mentions your business on Today and youd need 10 servers to handle the load but only for a few hours.

The better solution to both the resilience and scalability problem was cloud computing. Set up a server instance or two the little servers that run your applications on Amazon Web Services (AWS) or Google Cloud, and if one of the instances failed for some reason, it would automatically be restarted. Set up auto-scaling correctly and when Mr Humphrys causes the workload on your web server instances to rapidly rise, additional server instances are automatically started to share the workload. Later, as interest dies down, those additional instances are stopped, and you only pay for what you use. Perfect or is it?

Whilst the cloud solution is much more flexible than the traditional standalone server, there are still issues. Updating all the running cloud instances isnt straightforward. Developing for the cloud has challenges too: the laptop your developers are using may be similar to the cloud instance, but its not the same. If you commit to AWS, migrating to Google Cloud is a complex undertaking. And suppose, for whatever reason, you simply dont want to hand over your computing to Amazon, Google or Microsoft?

Containers have emerged as a means to wrap applications with all of their dependencies up into a single package that can be run anywhere. Containers, such as Docker, can run on your developers laptops in the same way as they run on your cloud instances, but managing a fleet of containers becomes increasingly challenging as the number of containers grows.

The answer is container orchestration. This is a significant shift in focus. Before, we made sure we had enough servers, be they physical or virtual, to ensure we could service the workload. Using the cloud providers autoscaling helped, but we were still dealing with instances. We had to configure load balancers, firewalls, data storage and more manually. With container orchestration, all of that (and much more) is taken care of. We specify the results we require and our container orchestration tools fulfil our requirements. We specify what we want done, rather than how we want it done.

Kubernetes (ku-ber-net-eez) is the leading container orchestration tool today, and it came from Google. If anyone knows how to run huge-scale IT infrastructures, Google does. The origin of Kubernetes is Borg, an internal Google project thats still used to run most of Googles applications including its search engine, Gmail, Google Maps and more. Borg was a secret until Google published a paper about it in 2015, but the paper made it very apparent that Borg was the principal inspiration behind Kubernetes.

Borg is a system that manages computational resources in Googles data centres and keeps Googles applications, both production and otherwise, running despite hardware failure, resource exhaustion or other issues occurring that might otherwise have caused an outage. It does this by carefully monitoring the thousands of nodes that make up a Borg cell and the containers running on them, and starting or stopping containers as required in response to problems or fluctuations in load.

Kubernetes itself was born out of Googles GIFEE (Googles Infrastructure For Everyone Else) initiative, and was designed to be a friendlier version of Borg that could be useful outside Google. It was donated to the Linux Foundation in 2015 through the formation of the Cloud Native Computing Foundation (CNCF).

Kubernetes provides a system whereby you declare your containerised applications and services, and it makes sure your applications run according to those declarations. If your programs require external resources, such as storage or load balancers, Kubernetes can provision those automatically. It can scale your applications up or down to keep up with changes in load, and can even scale your whole cluster when required. Your programs components dont even need to know where theyre running: Kubernetes provides internal naming services to applications so that they can connect to wp_mysql and be automatically connected to the correct resource.

The end result is a platform that can be used to run your applications on any infrastructure, from a single machine through an on-premise rack of systems to cloud-based fleets of virtual machines running on any major cloud provider, all using the same containers and configuration. Kubernetes is provider-agnostic: run it wherever you want.

Kubernetes is a powerful tool, and is necessarily complex. Before we get into an overview, we need to introduce some terms used within Kubernetes. Containers run single applications, as discussed above, and are grouped into pods. A pod is a group of closely linked containers that are deployed together on the same host and share some resources. The containers within a pod work as a team: theyll perform related functions, such as an application container and a logging container with specific settings for the application.

Four key Kubernetes components are the API Server, the Scheduler, the Controller Manager and a distributed configuration database called etcd. The API Server is at the heart of Kubernetes, and acts as the primary endpoint for all management requests. These may be generated by a variety of sources including other Kubernetes components, such as the scheduler, administrators via command-line or web-based dashboards, and containerised applications themselves. It validates requests and updates data stored in etcd.

The Scheduler determines which nodes the various pods will run on, taking into account constraints such as resource requirements, any hardware or software constraints, workload, deadlines and more.

The Controller Manager monitors the state of the cluster, and will try to start or stop pods as necessarily, via the API Server, to bring the cluster to the desired state. It also manages some internal connections and security features.

Each node runs a Kubelet process, which communicates with the API server and manages containers generally using Docker and Kube-Proxy, which handles network proxying and load balancing within the cluster.

The etcd distributed database system derives its name from the /etc folder on Linux systems, which is used to hold system configuration information, plus the suffix d, often used to denote a daemon process. The goals of etcd are to store key-value data in a distributed, consistent and fault-tolerant way.

The API server keeps all its state data in etcd and can run many instances concurrently. The scheduler and controller manager can only have one active instance but uses a lease system to determine which running instance is the master. All this means that Kubernetes can run as a Highly Available system with no single points of failure.

So how do we use those components in practice? What follows is an example of setting up a WordPress website using Kubernetes. If you wanted to do this for real, then youd probably use a predefined recipe called a helm chart. They are available for a number of common applications, but here well look at some of the steps necessary to get a WordPress site up and running on Kubernetes.

The first task is to define a password for MySQL:

kubectl will talk to the API Server, which will validate the command and then store the password in etcd. Our services are defined in YAML files, and now we need some persistent storage for the MySQL database.

The specification should be mostly self-explanatory. The name and labels fields are used to refer to this storage from other parts of Kubernetes, in this case our WordPress container.

Once weve defined the storage, we can define a MySQL instance, pointing it to the predefined storage. Thats followed by defining the database itself. We give that database a name and label for easy reference within Kubernetes.

Now we need another container to run WordPress. Part of the container deployment specification is:

The strategy type Recreate means that if any of the code comprising the application changes, then running instances will be deleted and recreated. Other options include being able to cycle new instances in and removing existing instances, one by one, enabling the service to continue running during deployment of an update. Finally, we declare a service for WordPress itself, comprising the PHP code and Apache. Part of the YAML file declaring this is:

Note the last line, defining service type as LoadBalancer. That instructs Kubernetes to make the service available outside of Kubernetes. Without that line, this would merely be an internal Kubernetes only service. And thats it. Kubernetes will now use those YAML files as a declaration of what is required, and will set up pods, connections, storage and so on as required to get the cluster into the desired state.

This has necessarily been only a high-level overview of Kubernetes, and many details and features of the system have been omitted. Weve glossed over autoscaling (both pods and the nodes that make up a cluster), cron jobs (starting containers according to a schedule), Ingress (HTTP load balancing, rewriting and SSL offloading), RBAC (role-based access controls), network policies (firewalling), and much more. Kubernetes is extremely flexible and extremely powerful: for any new IT infrastructure, it must be a serious contender.

If youre not familiar with Docker start here: https://docs.docker.com/get-started.

Theres an interactive, tutorial on deploying and scaling an app here: https://kubernetes.io/docs/tutorials/kubernetes-basics.

And see https://kubernetes.io/docs/setup/scratch for how to build a cluster.

You can play with a free Kubernetes cluster at https://tryk8s.com.

Finally, you can pore over a long, technical paper with an excellent overview of Googles use of Borg and how that influenced the design of Kubernetes here: https://storage.googleapis.com/pub-tools-public-publication-data/pdf/43438.pdf.

Find out more about Tiger Computing.

(Image credit: Future)

Enjoying what you're reading? Want more Linux and open source? We can deliver, literally! Subscribe to Linux Format today at a bargain price. You can get print issues, digital editions or why not both? We deliver to your door worldwide for a simple yearly fee. So make your life better and easier, subscribe now!

Here is the original post:
Kubernetes - taming the cloud - TechRadar

Read More..

State of the Cloud, November 2019 – Cloudwards

Hello and welcome to this latest State of the Cloud, our monthly column where we go over the biggest stories from the cloud and tech industries. October was an eventful month, but November moved at an almost breakneck speed, so strap in as we review all the goings on in our niche.

Subscribe to our monthly newsletter for updates on reviews, articles and investigations.

Of course, our usual suspects of Facebook and Libra, its supremely dodgy currency, will make an appearance, but were not just throwing our usual stones from our glass house. Also checking in are the streaming wars and some shorter stories. First, though, well be talking about a pretty serious security breach affecting one of our favorite VPN providers, NordVPN.

Early in the month, news came to light that a NordVPN server had been breached back in March 2018. The attack was possible due to an error made by the data center operator, which the VPN is no longer in business with, but no records were exposed.

All in all, its not the biggest deal, really, considering cybercrime can hit anybody and at any time, but it does raise the issue of transparency.

After all, why did we not find out till October 2019, almost 18 months later, that this attack had taken place? Anybody grounded in the realities of online business knows attacks happen and occasionally the people behind them will gain access, but as long as you encrypt all your data and, like in the case of NordVPN, dont keep user logs, theres no harm done, usually.

Though the service did quickly come out with a statement answering these questions, as well as an explanation of what exactly happened, we have decided to ding the service a little in our NordVPN review, if only just a few points. That said, we still have faith in NordVPN, though this episode does remind us all that nothing is completely safe online.

On a less serious note, the streaming wars heated up ahead of the launch of Disney+ in the U.S. on November 12, with launches all over the world following soon after. If youd like to know more about the service or would like to figure out how to access it from anywhere, check out our guide on how to watch Disney+.

The new behemoth on the scene offering the full Star Wars and Marvel universe experiences, as well as everything Disney has ever produced is a serious threat to existing channels. Somehow this hasnt deterred anybody from doing what theyre doing, though, and in some cases even launching new competitors.

However, much as we said in the August edition of this column, all that this is likely to do is make piracy more attractive. Disney has effectively stolen away a huge chunk of Netflixs library and is setting up a killer deal with Hulu that will package it with Disney+, but Netflix still has enough to offer viewers, so its unlikely people will give it up.

Having subscriptions to both Netflix and Disney+ will cost you roughly $20 per month, which should be doable for most. However, if you add any other entertainment subscriptions or start thinking in annual terms, then many consumers might decide its all costing too much.

While these large corporations are battling over market share, they might very well find the market is shrinking simply because people are busy torrenting.

Mark Zuckerberg became a meme after his testimony on Capitol Hill in May 2018, thanks to his robotic performance and evasive answers, not to mention his excessive drinking of water. Since then, the boy wonder hasnt gotten much in the way of media training, it seems, though several U.S. politicians have sharpened up their tech knowledge.

This was evident during The Zucks hearing on the Libra, his would-be digital currency. To cut a long story short, it seems that nobody within Facebook has any real idea of how Libra would work, what the risks are to consumers (and their privacy) and what exactly the company hopes to achieve with its introduction.

As you would expect, many backers of the Zuck Buck quickly abandoned ship after the massive public outcry following the hearing.

Adding to Facebooks woes and yes, you may picture us rubbing our hands in barely suppressed glee a massive class action lawsuit filed against it over its misuse of facial recognition technology has received the go-ahead from a judge, a federal one, no less.

The price tag for this could be as high as $35 billion, plus whatever court costs, potentially stymying Zuckerbergs steamroller for the foreseeable future. We can only hope.

To finish up our regular coverage, theres yet another chapter being written in the saga of whether or not Western countries should let Huawei build (part of) nascent 5G networks.

As we talked about in our June edition, the only companies capable right now of delivering on the superfast data connections are from the Middle Kingdom, and all those come with the very real threat of spying by the Chinese intelligence services.

Thus, the EU had a risk assessment commissioned on what we can and cannot expect when its member states let Huawei work on its 5G networks. Though its a long, nuanced report, what it boils down to is that the risks are manageable, provided that the right precautions are taken. In turn, Germany has decided to open the doors to Huawei, though not completely.

However, this completely ignores the problem that any time China is given any way in, it exploits this opening mercilessly; for examples, just see whats going on in its nearby waters or the brutal suppression of Hong Kong.

The problem lies in the fact that rational Western politicians see China as a reasonable partner, which it very much is not. Letting Huawei work on telecommunications is just sowing the wind.

To see the damage that could possibly be wrought by somebody with access to countrywide computer systems, look no further than Georgia the Caucasian nation, rather than the U.S. state. A massive cyberattack pretty much laid the entire country out, causing untold damage and likely seriously freaking people out.

However, according to at least one Chinese official, its actually the West that is to blame for all these problems, thanks to the Cold War mentality that hinders mutual trust in cyberspace. We have a feeling Chinese censorship and its human rights abuses might be partially to blame for that, too, though.

In better news, Google has banned a whole mess of payday loan apps from the Play Store, to the chagrin of said loan sharks. Anything that makes one of the most predatory industries upset is a good move in our book, so all praise to Mountain View in this case.

In other Google-related news, the company has unveiled a quantum computing breakthrough that, according to people a lot smarter than we are, is pretty great, but not as huge as announced. Well be honest, all we could think of is whether or not it could run The Witcher 3 on full settings on a big screen.

With that, well leave you for this month. November is shaping up to be another exciting month, and were already looking forward to telling you all about it in December. For now, we wish you luck with the darkening days in the Northern Hemisphere and much enjoyment of the spring in the Southern.

Subscribe to our monthly newsletter for updates on reviews, articles and investigations.

Did we miss any important news from October? Or did we wildly misinterpret any facts? Let us know in the comments below. Thank you for reading.

Read the original here:
State of the Cloud, November 2019 - Cloudwards

Read More..

How to ditch the cloud and move to do-it-yourself NAS instead – The Age

What is NAS?

NAS is essentially a bunch of hard drives that connects to your home network, powered and housed by a small computer, enabling a centralised file storage system you can access from anywhere.

Traditionally a NAS box required a bit of know-how to get running, but manufacturers have made great strides in this area to the point that almost anyone can set up a powerful network storage solution that is more capable and flexible than a cloud storage service.

Synology's DS218J is a powerful two-bay NAS box at an entry level price of around $230.

Since you're buying the NAS box and requisite hard drives outright, there's more of an upfront investment. But it works out cheaper in the long run as there are no ongoing monthly fees. Cloud storage is akin to renting a place for your data to live, whereas NAS is more like owning your own home, giving you complete control and ownership. Boxes are designed to run 24/7, but generally don't consume a lot of power.

A two-bay NAS box can be picked up from as low as $200. Filling those slots with two 1TB hard drives will set you back another $100, so in total you're looking at $300. By comparison, a Google One plan with 2TB storage will set you back $125 a year. I invested in a more expensive five bay Synology DS1019+, and filled up the hard drive slots as and when I needed more storage over time. More drive bays also give you better redundancy, as you can mirror data so you won't lose any if one or two drives fail.

Loading

Keep in mind that NAS boxes can do more than just store and share your photos. The likes of Synology and QNAP have an extensive app ecosystem that extends the functionality well beyond the bounds of traditional network storage.

I use mine as a media server so the family can easily stream movies and music stored on the NAS to any connected smartphone, tablet, PC or streaming box, in addition to serving as a PVR for recording major sporting events on free-to-air television. I also use it to drive and monitor my home security cameras.

There are a number of companies that make NAS boxes, including QNAP, Western Digital and Seagate, but Synology's DiskStation line is far and away the best in the industry when it comes to ease of use, stability and features.

For example making your NAS accessible from outside of your home network usually involves setting up port forwarding rules or other complicated network settings. But Synology's QuickConnect feature bypasses this by allowing you to assign a simple URL or ID for access.

The DS1019+ is a five-bay NAS box that supports 4K transcoding expansion bays for even more storage, at around $1000.

It's also the only NAS system that can match Google Photos in terms of smarts. Synology's Moments app, which runs on top of the company's Diskstation Manager operating system, analyses all your photos and puts them into sensible albums for you, making it much easier to find the photo you're looking for.

It uses facial recognition to group photos with similar faces, and scene recognition so you can search based on things that are in the picture.

Loading

Unlike Google Photos, Synology does all this without ever collecting any user data or sending a single photo to the public cloud.

You can also set the app to automatically upload any new photos from your phone to your NAS.

Another strong point for NAS compared to cloud storage services is speed. Cloud services are limited by your internet speeds and the bandwidth of their servers, whereas NAS utilises the speed of your local home network which is significantly faster.

Of course it's always wise to keep a backup offsite of all your important files in case there's a fire or burglary. Synology has multiple options for doing this, backing up data stored on the NAS to a public cloud service like Google Drive, OneDrive or Dropbox.

The benefit here is that Synology will encrypt your data before it is uploaded, so your data can't be trawled and won't be compromised if the cloud service is hacked or breached.

Some NAS boxes allow you to sync an encrypted backup of your storage to the cloud.

Synology also offers its own private cloud option called Synology C2 Backup, with the basic plan costing between $16 and $100 a year depending on your needs.

Another option is to invest in a secondary Synology NAS offsite and have files synchronise over the internet. I personally go the manual route; plugging a USB drive into the NAS on a monthly basis to back up my most precious data, namely my collection of family photos and videos.

Krishan is a multi-award-winning Australian technology journalist.

Here is the original post:
How to ditch the cloud and move to do-it-yourself NAS instead - The Age

Read More..

Inspur steps up with Innovative Liquid Cooling Technology at SC19 – insideHPC

Inspur is showcasing new HPC systems with Natural Circulation Evaporative Cooling technology this week at SC19. Inspur combines high-density computing servers with natural circulation evaporative cooling technology, which is more reliable, energy-saving, and easier to deploy than other liquid cooling solutions.

Currently, HPC is in urgent need of cooling technologies with high efficiency and lower energy consumption. However, the cost, safety, deployment and maintenance challenges of cooling solutions are big concerns of users. Liquid cooling technologies, compared to traditional air cooling counterparts, boast prominent advantages in heat dissipation efficiency, energy utilization, and other aspects, and have seen rapid growth in recent years. Nevertheless, liquid cooling technologies are still subject to challenges. For example, immersion and spray-type liquid cooling technologies, despite their more efficient heat dissipation, require to have the IT components continuously observed to maintain the functionality and reliability because of direct contact with the coolant. Moreover, excessive usage of coolant pushes the limit on machine room load-bearing capacity, and raises operation and maintenance costs. Plate-type water cooling technology, though not exposed to heating elements directly, uses uninsulated water as the coolant which, once leaked, will cause lethal damage to HPC systems, giving rise to safety hazards.

Inspur collaborated with the Institute of Electrical Engineering of Chinese Academy of Sciences (IEECAS), combining Inspurs leading supercomputing servers with IEECASs natural circulation evaporative cooling technology to achieve an efficient, reliable and energy saving liquid cooling HPC system. The system is equipped with Inspurs high-density server i24 which can support 4 high-performance two-socket computing nodes in 2U and the natural circulation evaporative cooling suite developed by IEECAS. The natural circulation evaporation method requires no circulating pumps which are necessary in traditional plate-type water cooling but vulnerable and energy consuming and enables automatic control over condenser fans, eliminating manual operation for over 90% of the time. This further reduces cooling overheads and lowers the PUE values of data centers to below 1.1 for green and energy-efficient operation. The non-corrosive insulating cooling liquid protects IT devices from damage in the event of leakages, greatly improving safety. In addition, the entire cooling system is compact in size and easy to deploy and maintain, with less demanding requirements on machine rooms.

The cooling system has already been successfully deployed in a large science project and is performing as expected.

Inspur is a leading provider of data center infrastructure, cloud computing, and AI solutions, ranking among the worlds top 3 server manufacturers. Through engineering and innovation, Inspur delivers cutting-edge computing hardware design and extensive product offerings to address important technology arenas like open computing, cloud data center, AI and deep learning. Performance-optimized and purpose-built, our world-class solutions empower customers to tackle specific workloads and real-world challenges.

See our complete coverage of SC19

Check out our insideHPC Events Calendar

See the article here:
Inspur steps up with Innovative Liquid Cooling Technology at SC19 - insideHPC

Read More..

Evaluating Internet Isolation Clouds: Must-Have Features – Security Boulevard

Not all Internet isolation solutions are created equal. Just ask our customers. One of them learned the hard way.

With millions of customers worldwide and trillions of dollars in assets, a leading global investment firm was a high-profile target for cybercriminals. A layered defense protected the organization against a broad variety of threats, but email phishing attacks were becoming an increasingly serious threat. Something had to be done.

To combat email threats, the organization deployed multiple layers of security. The architecture included cloud and on-premises versions of anti-spam, anti-virus, data security, encryption, and sandboxing solutions, but spearphishing attacks and drive-by malware exploits were still a significant risk. It would take only one successful attack to cause billions of dollars worth of damage.

Knowing that this was an unacceptable risk, the organization worked with Menlo Security to implement a Zero Trust Internet cybersecurity strategy. Today, all of the organizations web sessions pass through the Menlo Security Internet Isolation Cloud, isolating all web traffic. When users do click on malicious links, all sites are already safely isolated and have input-field restrictions. By isolating all email links and attachments, Menlo Security Email Isolation protects the firm against credential theft, while eliminating 100 percent of drive-by malware exploits.

Here are 5 out of 10 must-have features to look for when evaluating an Internet isolation cloud, you will need to download the ebook to see the others:

Web Isolation Technology

Protection Against Phishing

Should include and integrate email security, including automatically isolating the links and attachments from email

Prevents users from filling out suspicious web form

Native-Like User Experience

Advanced Threat Protection

Should assume that all content originating from the Internet is risky, including content that was previously benign

Provides 100 percent protection against all web- and document-borne threats, including zero-day malware

Additional Cybersecurity Features

Should go beyond just isolation to provide AV, URL filtering, DLP, reporting, policy and authentication, encrypted traffic management, cloud access security broker, and email link and attachment protection

Dont wait until your organization suffers a devastating breach or until you reach the tipping point when cybersecurity becomes cost prohibitive as a result of inefficiencies. Look for an Internet isolation cloud that has these 10 features. You wont be sorry.

Go here to see the original:
Evaluating Internet Isolation Clouds: Must-Have Features - Security Boulevard

Read More..

Perimeter 81 Announces $10 Million Funding Round to Expand its Network as a Service Platform; Partners with SonicWall to Add Unified Security Services…

The companies will integrate SonicWall's advanced security solutions and Perimeter 81's matured and innovative Zero Trust Network as a Service solution. Both company's platforms will offer an easy-to-use "Secure Network as a Service" solution that provides Zero Trust access to internal resources, user and branch internet security, branch interconnectivity and endpoint security in one place.

With cloud and mobility solutions changing the way we work, we are seeing an increasing number of employees working on the go. This is driving a need to shift from a dated site-centric security model to a more cloud user-friendly, service-based platform.Perimeter 81 is one of the only vendors offering a comprehensive set of converged secure network capabilities, delivered and managed from its multi-tenant cloud.

This funding round is the latest move in a year of exponential growth for Perimeter 81. Launched in February 2018, the company has already seen 400% year over year revenue growth, expanding from 180 businesses using Perimeter 81 to 500 in the past eleven months alone. More than 81% of Perimeter 81's customers use it as their secure corporate network and not only as Zero Trust Access, replacing the need for a VPN.

"As we see greater numbers of companies, from SMBs to enterprises and governments, making the move to a perimeter-less organization largely supported by cloud services, Perimeter 81 has seen tremendous growth," said Amit Bareket, Co-Founder and CEO of Perimeter 81. "When we created our Network as a Service solution, we made it possible to easily offer security solutions on top of our existing platform. We are excited to partner with SonicWall in order to provide the most advanced Secure Network as a Service available today."

"We are excited to partner with Perimeter 81. This relationship will allow us to provide a wide range of businesses, from SMBs to Fortune 500s and governments, with our award-winning Capture Cloud Platform and real-time breach detection and prevention solutions, while allowing them to adopt a Zero Trust Security architecture that delivers tremendous efficacy in securing the modern organization," said SonicWall President and CEO Bill Conner. "Going forward, our customers and partners have choices between on-premise, cloud, and now, Secure Network as a Service all with integrated management within Capture Security Services. We believe this will ensure that our customers continue to fearless."

About Perimeter 81

Perimeter 81 is a Zero Trust Secure Network as a Service that is simplifying network security for the modern and distributed workforce. Based in Tel Aviv, the heart of the startup nation and a global hub for innovative technology development, Perimeter 81 was founded by two IDF elite intelligence unit alumni, CEO Amit Bareket and CPO Sagi Gidali. The team of security as a service experts comes together every day to deliver a truly innovative, world-class network security service. Perimeter 81's clients range from SMB to include Fortune 500 businesses and industry leaders across a wide range of sectors, and its partners are among the world's foremost integrators, managed service providers and channel resellers. Earlier this year, Gartner selected Perimeter 81 as a "Cool Vendor in Network and Cyber-Physical Systems Security. For more information, visit http://www.perimeter81.com

About SonicWall

SonicWall has been fighting the cybercriminal industry for over 28 years defending small and medium businesses, enterprises and government agencies worldwide. Backed by research from SonicWall Capture Labs, our award-winning, real-time breach detection and prevention solutions secure more than a million networks, and their emails, applications and data, in over 215 countries and territories. These organizations run more effectively and fear less about security. For more information, visit http://www.sonicwall.com or follow us on Twitter, LinkedIn, Facebookand Instagram.

SOURCE Perimeter 81

Perimeter 81 | Homepage

Read this article:
Perimeter 81 Announces $10 Million Funding Round to Expand its Network as a Service Platform; Partners with SonicWall to Add Unified Security Services...

Read More..

1.19 billion confidential medical images available on the internet – Help Net Security

1.19 billion confidential medical images are now freely available on the internet, according to Greenbones research into the security of Picture Archiving and Communication Systems (PACS) servers used by health providers across the world to store images of X-rays as well as CT, MRI and other medical scans.

Thats a 60% increase from the finding between July and September 2019, and includes details of patient names, reason for examination, date of birth, and ID cards in some cases.

Amongst the 786 million medical images identified in the US, which had the largest increase in new data sets discovered, Social Security Numbers were included on some of the images, as well as some sets which listed details pertaining to military personnel IDs from the Department of Defense.

Overall, 129 new easily accessible archiving systems and data from nine additional countries have been discovered. Also, the number of images freely available on the internet had increased most significantly in the US, India, South Africa, Brazil and Ecuador.

Proper controls, such as HIPAA in the US, were largely missing. In total, the number of data records which are accessible online without any level of protection has doubled, from 4.4 million to 9 million, and the number of images now accessible or easily downloadable via the internet is 370 million.

Conversely, 172 PACS servers, including all systems from 11 countries including the UK, Germany, Thailand and Venezuela, had in fact been taken completely offline and the patient data was no longer accessible via the internet.

Dirk Schrader, cyber resilience architect at Greenbone Networks said: Whilst some countries have taken swift action to address the situation and have removed all accessible data from the internet, the problem of unprotected PACS systems across the globe only seems to be getting worse. In the US especially, sensitive patient information appears to be free-flowing and is a data privacy disaster waiting to happen.

When we carried out this second review, we didnt expect to see more data than before and certainly not to have continued access to the ones we had already identified. There certainly is some hope in the fact that a number of countries have managed to get their systems off the internet so quickly, but there is much more work to be done.

Read more:
1.19 billion confidential medical images available on the internet - Help Net Security

Read More..