Category Archives: Cloud Servers
Internet of Things has started to impact every aspect of our daily lives. Our appliances, cars, gadgets, communication devices, tools, and even some of our clothing have become nodes on the internet. By 2020, as many as 50 billion devices will be connected so there will be skyrocketing growth of traffic generated by devices at the edge of the network posing a monumental challenge to our networks and central cloud computing. Thankfully, we can take advantage of the ever-increasing computing capabilities of edge devices to turn them into cloud servers and extend central-cloud capabilities to the edge. Edge cloud to central cloud is like WiFi to Cellular Communication. Just as WiFi carries most of the global wireless traffic today, edge devices will soon manage most of the cloud computing burden.
According to Gartners 2017 Hype Cycle for Emerging Technologies, edge cloud computing is on the brink of becoming an innovation trigger. Microchips and sensors continue to become embedded in everyday objects, making edge cloud computing an immense opportunity. There are billions of potential edge servers today and there will be tens of billions more tomorrow.
An illustrative example: self-driving cars
Self-driving cars have many potential benefits: optimized traffic flow, improved fuel efficiency, reduced accidents, and drivers with more free time. However, a big challenge for self-driving cars is to find a cost-effective way to process the vast amounts of data that they generate. On average, every self-driving car generates approximately one GByte/sec of data which is orders of magnitude more data than the capacity of a 4G base station and an order of magnitude large than a 5G base station. There are no networks in the foreseeable future that can be used to send all this data back to the central cloud for processing. Most of this data needs to be processed locally and only a minimal set of processed data should be sent back to the central cloud for global coordination. Moreover, in order to meet latency requirements to take agile decisions, self-driving cars should communicate in the fastest way possible. This demands instantaneous processing of information and when possible peer to peer communication. This is where distributed edge cloud computing comes into play, transforming cars to data centers on wheels where most of the communication and processing is performed as close as possible to the edge.
In a distributed edge cloud environment, every car can be a set of microservices that can sense other cars not only because of all the sensors but also because their microservices can communicate with microservices of other cars. To make this work, microservices (within a car and between cars) need to form ad-hoc clusters based on scopes such as proximity, network or account affinity. This way, cars can process the terabytes of data as quickly and as efficiently as possible leveraging not only the processing power at the central cloud but also their own collective computing, storage and memory resources in a collaborative fashion with other cars on the road.
Edge cloud computing is the next major computing revolution
Throughout the last few decades, computing has experienced different cycles shifting between centralized and distributed. In the early days, computing was centralized with mainframe computing. In the mid-80s, with the advent of personal computers the industry shifted to distributed computing. In the last decade, we witnessed the move to centralized cloud computing. Many falsely predicted that this was the holy grail of computing and we would move to an era of thin clients where devices would be dumb screens and input devices and all the processing would be performed in data centres in the cloud. This made sense for some applications such as music or video streaming or hosting some software applications. In all these cases, edge devices do not generate too much data and are mostly passive receivers of information.
In the last few years, two major socio-technical trends have contributed to a fundamental change in production and consumption of data.
First, thanks to mobile internet and social media, ordinary people generate massive amounts of data turning them from mere consumers to consumers and producers. For example, today close to 500 million photos are uploaded on Facebook and Instagram and roughly 500 thousand hours of video is uploaded to YouTube daily; this is more than what the three major US networks generate in content over two years!!! This is not a consumer phenomenon but also applies to enterprises; for instance, more than 80% of businesses have started to leverage user-generated content in their marketing efforts.
Second, we have the rapid growth of IoT devices where many new edge devices produce valuable data. There are already 20 billion connected devices, 2.5 billion of which are B2B IoT devices. Over 300M wearable devices were sold in 2017 alone. Many of these devices generate small amounts of data but many generate massive amounts of data; for example, when video meets IoT. We apply machine-learning algorithms to video feeds allowing cameras to recognize people, objects, and situations automatically. There will be phenomenal growth in AR/VR in the gaming industry, and even the enterprise starting with creative applications and quickly moving to brick and mortar industries and manufacturing. Robots will also be producers of massive amounts of data at the edge.
Clearly, we are amid an explosion of data generated at the edge and the tsunami is yet to come. The question is can our communication networks scale to cope with the data generated at the edge? To try to answer this, we can look at two predictors: the Moores law in computing and its equivalent in network bandwidth. History has shown that computing power roughly doubles every 18 months (or hundred times every decade) whereas network bandwidth grows about 50 times every decade. In other words, even if the number of devices do not grow (which they clearly will), the communication network will be the bottleneck for the growth of IoT.
Setting bandwidth aside, many IoT applications such as self-driving cars or tactile control communications in various industries require low latency response. In this case, even if the network capacity is miraculously increased to cope with the data, laws of physics inhibit remote processing of data in the central cloud due to large latencies in the long-haul transmission of data.
So, what is the solution? How can we cope with the explosion of data at the edge and strict latency requirements of some IoT applications? The answer is distributed edge cloud computing. Edge cloud computing means that any device (or node) becomes a cloud server. As much as possible, the data is processed at the edge of a network, as close to the originating source as possible, instead of processing everything in the central cloud. This approach is faster, more efficient, and scalable: data can be immediately analysed and put into action overcoming bandwidth limitations and latency constraints on the network. Edge cloud computing is essential to meet stringent requirements on bandwidth and latency and at the same time minimizes power consumption and infrastructure costs.
Edge cloud computing is a paradigm shift that enables every device, appliance, or gadget to communicate and share resources making them part of the solution for scaling of IoT. It allows drones and robots to harness their collective resources in industries such as manufacturing, oil and gas, agriculture or mining, delivering real-time data and improving business efficiency. This new computing model will revolutionize the world in ways that we may not be able to predict at this moment.
The great news is that the technology is ready for developers today. mimik has developed a fully distributed edge cloud platform that extends central cloud to the edge: mimik arms developers with a platform to unleash the power of edge devices. mimik SDK solves many of the current challenges that centralized cloud computing alone cannot address. We extend the power of the central cloud to the edge and boost its reach so that bandwidth, latency constraints, and infrastructure cost do not become the bottleneck for the healthy and scalable growth of IoT.
We need a paradigm shift that transforms tens of billions of devices from a challenge to an opportunity. IoT requires a revolution in computing that unlocks the power of connected devices. Distributed edge cloud is the ideal solution to harness computing resources at the edge, unlocking the promise of a smart connected world that will bring massive efficiencies to enterprises and digital freedom to consumers.
Siavash Alamouti, CEO at mimik
Image Credit: Jamesteohart / Shutterstock
We generally think of a transformation to a hybrid infrastructure as one where you’re going from a completely private setup to one that spans the public cloud and your private installation. But what if you started life as a small company with your systems entirely in the cloud? It’s not an unusual approach, as running up your initial services in the cloud is straightforward and avoids a big capital outlay. As a company grows it’s understandable that it might want to take on a private data centre, build an in-house support team and evolve to a two-site setup.
Step one is to consider why you’re bothering with an on-premises setup instead of a second cloud instance. The answer will generally be that you want something that’s closer to your office, with a potential performance improvement gained from such proximity. And that’s fine what matters is that you’ve considered the options before deciding which way to go.
The next step is to think about where you’ll host your private data centre. As you’re already in the cloud, you have the opportunity to pick a data centre that’s close (electronically speaking) to the cloud centre you’re in. For example, you’re probably aware that AWS provides a Direct Connect facility that lets you hook straight into their infrastructure rather than accessing your cloud world over the internet. Check out the locations and you’ll see that the connectivity’s hosted at 51 well-known locations Equinix in London, for example, or TierPoint in Seattle. Connectivity between your public and private components with a latency of just a few milliseconds is an attractive concept if you’re looking for high availability with seamless failover.
Next, you’ll need to think about the platform you’re using. Most of the time you’ll have used one or more of your cloud provider’s standard operating system templates, so it makes sense to run your local stuff on the same operating system flavour if you can. And of course you should use the same CPU architecture where you can too, so you can be assured that your apps will be portable.
So you’ve sorted the platform. Now you need to decide whether the on-premises setup is to be your primary or secondary installation. If it’s to be a secondary setup you should have a relatively straightforward job of adding new system and application-level components in as secondaries to your cloud-based apps.
If you decide to flip things around you’ll have a more involved task of be shifting the primary apps over and redeploying the cloud setup as the secondary installation. Either way the happy news is that you’ve already gone through the non-trivial task of providing your office users with connectivity to the cloud installation, so hooking things up so they’re able to get to the private data centre, regardless of whether it’s the primary or the secondary, should be easier.
One further consideration with the choice of primary and secondary installations is the cost of data transfer. Shifting data out of a commercial cloud setup has a cost associated with it. Not a vast cost, I’ll grant you, but one that you do need to keep an eye on. Using Amazon as an example, moving a terabyte per month over the internet from the cloud setup to your private installation will cost you $90. That’s $900 for 10TB, or $7,800 for 100TB; even though the per-gigabyte cost tapers down, it doesn’t ever tail off at zero. What does this mean? Easy: if the cloud setup is the primary and it’s replicating application data to the private secondary, you’re paying a chunk of cash for it to do so.
While we’re on the subject of data transfer, you also need to figure out how you’re going to do it. In these modern times, it’s a relative doddle to set up the major cloud providers’ storage instances so you can access them externally via standard protocols such as NFS. Alternatively you can look to the major storage vendors, who will sell you a funky gateway to install in your private data centre and handle the cloud magic for you.
The next consideration is licensing, and there are two aspects here. First is the basic fact that you’ll need to buy operating system and/or application licences for your private setup sounds obvious but you may not ever have had to consider this if you were using a pay-as-you-go model with pre-configured cloud app servers. Second is that if you want to go for a clustered or active/passive application setup, you may need to revisit the versions you use on the cloud servers as well as buying licences for your private setup. Take SQL Server, for example: if you’re running Standard Edition you can implement basic two-node high availability, but if you want something more advanced you’ll need to upgrade to Enterprise Edition. Same with Oracle: if you want to enable Data Guard between sites that’ll need Enterprise Edition too.
Lastly, but by no means least, is your internal support team. They’ve probably spent a number of years fettling your cloud installation and fixing stuff when it broke, but their skillset will be at worst lacking and at best out of date when it comes to hosting, networking, hardware and hypervisor support.
Be prepared to invest in training so that you can be confident that the new kit you’re acquiring for your private data centre is properly supportable and hence properly supported. Yes, your typical infrastructure is easier to put together than it was a few years ago, but that doesn’t mean it’s trivial. And if you’re virtualising your private data centre which you should getting the hypervisor layer running and optimised will take time, effort and skill.
Going from a cloud-centric setup to a hybrid infrastructure isn’t rocket science, then which is no great surprise as any problem’s tractable if you design, plan and implement the solution properly. But going from cloud to hybrid has some differences from going from private to hybrid.
So you just need to think a bit before you do it.
Sponsored: The Joy and Pain of Buying IT – Have Your Say
By PATRICK LANG’ATMore by this Author
The Raila Odinga-led National Super Alliance (Nasa) has written to the electoral agency with 25 new demands that they say should be fulfilled in the October 17 fresh poll.
In a letter signed by Mr Odingas chief agent and Nasa co-principal Musalia Mudavadi, the opposition has demanded a full audit of the elections technology, full access to the servers, change of ballot printing firm, and the gazettement of new 290 constituency returning officers.
It is therefore inappropriate and foolhardy for the IEBC to embark on the planning of the fresh election without full compliance with the Orders of the Supreme Court in the redeployment of technology in the fresh presidential election. We therefore demand a full audit of technology in use in full compliance with the law, Mr Mudavadi said in the four-page letter.
The Nasa team has not only opposed the Independent Electoral and Boundaries Commission date for the repeat poll, it has also questioned why it narrowed the number of candidates to only Mr Odinga and President Uhuru Kenyatta.
Your interpretation that the election scheduled on the 17th October 2017 shall be in the style of a run-off contest of only two candidates is erroneous and unconstitutional. We also take reservation that you have not consulted the parties involved before making a determination on the date of the said election, Mr Mudavadi told IEBC Chairman Wafula Chebukati.
In the audit, Nasa has demanded a scrutiny they say should have full information on the ICT infrastructure, list of support partners and their respective Service Level Agreements, a detail of the firewall configuration, including ports configuration, as well as disclosure of all database transaction logs.
The team also wants a physical view and inspection of the IEBC servers, portal access to the cloud servers and IP addresses of all 20 servers; full access and copy of all servers and databases used by the IEBC; GPS coordinates of KIEMS; and Telkom and network structure with all service providers.
The opposition has also demanded the removal of senior personnel at the IEBC secretariat including Chief Executive Ezra Chiloba, his deputy Betty Nyabuto, James Muhati, the ICT director, Ms Immaculate Kassait, director of voter registration, Ms Praxedes Tororey, the head of legal team, with Mr Moses Kipkosgey being added to the earlier list.
Independence of the IEBC is not negotiable. Nasa coalition demands that fresh election should be administered by professional and non-partisan officials, Mr Mudavadi said.
“We demand that officials who are partisan or perceived as such should step aside and or be suspended during the planning and execution of the fresh election.”
Further, the coalition demanded a full audit of the Sh3.8 billion 45,000 voter identification and results transmission kits that were provided by French-based Safran Morpho.
We demand that Safran and Oracle provide full implementation information in relation to their involvement in the General Election held on 8th August 2017, said Mr Mudavadi.
The team has also demanded a review of the voter register, and the 40,883 polling stations.
To enhance transparency, Nasa said the following specific demands must be met:
Appoint and gazette returning officers not among the Constituency Election Coordinators in consultations with political parties and candidates.
Establishment of a technical monitoring committee with representatives of the main political parties, coalitions or candidates to oversee implementation of the technology in use.
Stop use of Al-Ghurair to print ballot papers and results declaration forms
All Forms 34Bs should be pre-printed indicating the names of polling stations in the constituency and names of candidates
Elections results to be announced at the Constituency level. Results sent electronically must be accompanied by corresponding statutory result declaration forms
Candidates agents should be part of receiving teams at the constituency and national tallying centers, and be allowed to confirm entries before transmission
Establish defined roles of the security agencies and eliminate undue influence by the provincial administration and other public officials
Jubilee Party chose the October 17 date, claim Nasa leaders.
Follow this link:
Nasa: Our demands for repeat presidential election – Daily Nation
NEW DELHI: Xiaomi said it was open to moving its servers to India subject to its cloud service provider partner setting up base in the country, amid increased government efforts to protect user data on mobile phones.
All our servers are sitting on AWS (Amazon Web Services) in Singapore and US. If AWS moves to India, we would be happy to work with them, Manu Kumar Jain, managing director of India operations, told ET, becoming the first overseas company to openly offer to move its servers to India.
Handset companies typically dont store data on their own servers but instead lease space on third party cloud service providers such as AWS, Microsoft and Google. While AWS and Microsoft have already set up their centres in India, Google has also announced setting up the same in the country to cater to a larger number of customers, especially those in the government or financial services industry since regulations in those sectors dont permit data to be transmitted outside the country.
We last evaluated this about 2-3 years ago when we were moving our servers (from China). At that time there was no significant presence (of AWS) and it was much more difficult to have it here, Jain said, when asked whether the company would move or add its servers in India, which appears to be the larger aim for the government intending to secure data. Jain did not say whether the company was already in talks with Amazon to move its servers to India.
He though added that from an internet speed perspective, the connectivity between India and Singapore was one of the best. We moved and thought it was pretty good. But if someone, AWS or equivalent, were to set up servers here (in India), we would be happy to work with them, he added.
The company, which sells about 4-5 million smartphones a quarter, said its devices were super-secure, no data was taken without use consent and the data that is taken is encrypted to the highest degree and that it cannot be decrypted even if the data is stolen.
Xiaomis views come at a time when the government is taking up security of phones with all companies, including Apple and Samsung, and scrutinising protection levels that all handset makers a large majority of which are Chinese were providing in India.
Another Chinese brand One-Plus has also said that it is prepared to respond to Indias data security and privacy concerns, since it sells the same smartphones in the US and Europe, where these concerns are already addressed.
Currently, we have not received direct request or requirement to set up servers or cloud storage in India. We are trying to get more clarity on that, One-Plus CEO Peter Lau told ET.
Amid the recent India-China standoff at Dokalam which has since been resolved the IT and electronics ministry has asked over 30 smartphone companies for protocols used by them to ensure the security of mobile phones in the country.
While the government is evaluating responses on security preparedness, it may well ask all those selling devices in India to have locally based servers.
Officials in the ministry said while the issue of apps sweeping up excessive user data was worrying, the broader issue remained that of the security of information that could be going to third parties outside the country, especially to China.
VMware officially lands on AWS cloud with new management and security features – SiliconANGLE News (blog)
Nearly a year after signing a landmark deal to bring its software-defined data center technology to the Amazon Web Services Inc. cloud, VMware Inc. kicked off its VMworld conference in Las Vegas today with the news that VMware Cloud on AWS is now generally available.
The service essentially enables the vast majority of companies that use VMware inside their data centers to use VMware software, which allowsdifferent operating systems and multiple applications to run on the same physical computer,with AWS services as well.
To date, companies have had difficulty moving workloads to Amazons cloud to take advantage of the clouds more flexible and lower-cost computing and storage services because many of their applications depended on VMware software that only ran on computers in company data centers. That presented customers of each provider with a tough choice: Use VMware technology it built its core applications on, but with none of the cost and flexibility of cloud computing, or use Amazons cloud, but not with the VMware software their data centers are built on.
They hated this binary decision that we were forcing on them, AWS Chief Executive Andy Jassy (pictured, right) said during an appearance this morning at VMworld with VMware CEO Pat Gelsinger (left). Now, the executives said, customers can more easily use so-called hybrid cloud services that use both on-premises software and hardware and cloud services as needed.
If this fully works, CIOs have no excuse in regard to moving VMWare loads to the cloud, said Holger Mueller, vice president and principal analyst at Constellation Research. But lets see if this works.
VMware, part of Dell Technologies Inc.s constellation of companies that also includes storage supplier Dell EMC, also announced a raft of services for the VMware Cloud today. Initially, VMware Cloud is available in the AWS U.S. West region, but other regions will be added throughout 2018.VMware said the integration will enable customers to run applications across operationally consistent vSphere-based private, public and hybrid cloud environments with the option of expanding to AWS elastic or bare-metal infrastructure.
When the AWS-VMware deal was announced last October, it was apparent that it could reset the competitive environment in computing, in particular presenting new challenges for IBM Corp., which had signed a deal with VMware earlier in 2016, Google Inc.s cloud platform and Microsoft Corp., whose No. 2-ranked Azure public cloud had claimed the lead in hybrid cloud computing.
The arrangement with AWS offers some benefits for VMware, including a connection to the leading public cloud provider that its customers have been clamoring for. When your own cloud fails, you need to join the ones that work, Mueller told SiliconANGLE. VMware now focuses on add-on software, such as application security.
But it also means AWS could steal some of VMwares customers ultimately, if it results in what Dave Vellante, chief analyst at SiliconANGLE Medias Wikibon, has called a potential one-way trip to Amazon cloudville.’ Moreover, said Mueller, the arrangement doesnt help Dell sell more servers into an on-premises data center.
As for Amazon, Mueller said, AWS needs a piece of the on-premises enterprise load and this is the way. He added that the fact that AWS is offering to host VMware instances on so-called bare-metal servers, those with no operating software installed on them, indicates how much it needs VMwares help to reach large enterprise customers, since AWS had generally eschewed bare-metal arrangements.
The offering will be delivered, sold and supported by VMware as an on-demand service. Its powered by VMware Cloud Foundation, a software-defined data center platform that includes vSphere, VMware VSAN and VMware NSX virtualization technologies managed by VMware vCenter. The initial set of cloud services includes six modules:
Discovery centralizes inventory information and cloud accounts across AWS, Microsoft Azure and VMware clouds, making it easier for information technology departments to search for and identify workloads. Administrators can group cloud resources even if they span multiple clouds. Built-in search and filters enables administrators to filter resources based upon cloud attributes.
AppDefense protects applications by embedding application control and threat detection and response capabilities into vSphere-based environments. Its tightly integrated with the NSX networking platform, and operates within the vSphere hypervisor to create a knowledge base of the correct state and behavior of each endpoint for change detection.
Cost Insight helps organizations analyze their cloud spending and identify savings opportunities. It provides detailed visibility into public and private cloud costs on AWS, Azure and VMware environments and enables drill-down to identify cost drivers. Cost Insight also identifies stopped virtual machines and associated storage resources across public and private clouds to reduce waste.
Network Insight analyzes application traffic flows between different tiers, virtual and physical network layers and public and private clouds. This has application security and load balancing applications, and makes it easier for cloud administrators to manage and troubleshoot large-scale NSX deployments.
NSX Cloud provides a single management console and common application program interface for monitoring and securing applications that span multiple private and public clouds. It features a micro-segmentation security policy that can be defined once and applied to application workloads running anywhere.
Wavefront is a metrics monitoring and analytics platform that gives developers insight into the performance of highly-distributed cloud-native services to detect performance anomalies while enabling high availability. Operating on what VMware said is a massive scale, Wavefront gives DevOps teams instant visualization of millions of data points per second. This helps resolve bottlenecks more efficiently and proactively.
VMware also said its expanding Cloud Foundations scope with new partner offerings. They include support from CenturyLink Inc., Rackspace Inc. and Fujitsu Ltd. New hardware platforms that support Cloud Foundation include Dell EMCs VxRack SDDC, Hitachi Data Systems Corp.s UCP-RS, Fujitsu Primeflex and Quanta Cloud Technology LLCs QxStack.
VMwares shares closed up nearly 2 percent today, to about $104.68 a share, on a relatively flat day for the overall market.
With reporting from Robert Hof
(* Disclosure: SiliconANGLE Medias video unit, theCUBE, is a paid media partner at VMworld. Stories on SiliconANGLE are written independently of coverage on theCUBE. Sponsors have no editorial influence on content on SiliconANGLE or theCUBE.)
VMworld 2017 VMware CEO Pat Gelsinger last week introduced the company’s second quarter results by saying the company has embarked on a multi-year journey from a compute virtualization company to offer a broad portfolio of products driving efficiency and digital transformation.
And today at VMworld the company began to explain what that mouthful of jargon meant: a strategy to put the company at the center of multi-cloud management.
The clearest expression of Gelsinger’s words is its half-dozen new software-as-a service offerings, namely:
All six are subscription services, accessible through existing VMware accounts. And all six are new stuff for your VMware account manager, or channel partner, to suggest. If you’re one of the few who resisted the company’s No Naked vSphere push, VMware’s going to come at you again, this time as a software-as-a-service vendor.
The Register expects the company will come hardest with AppDefense, because it’s created a new business unit to back a product it feels is genuinely new to offer. Most security is about finding bad, we are about ensuring good, says Tom Corn, senior veep of the Security Product group at VMware.
The Register revealed the basics of AppDefense well before its announcement. We had to wait for today to learn that it can build its whitelist of acceptable VM behaviour by interacting with either vCenter or automated provisioning tools like Jenkins or MAVEN. Linking with those tools is an effort to make AppDefense offer something to DevOps practitioners. It’s also trying to impress line-of-business types by offering them a mobile app that alerts them when applications misbehave, so that all stakeholders can participate in decisions about how to respond.
AppDefense will be sold as SaaS or on-premises software. Either way, it should do well: security types The Register’s virtualization desk have spoken to feel Virtzilla is onto something here!
VMware’s favourite news from this year’s event is that the company’s deal with Amazon Web Services has come to fruition. AWS now hosts servers running Cloud Foundation, the bundle of vSphere, VSAN, NSX and vCenter that is intended to mirror on-premises implementations.
It’s all available as of today, to run in AWS alone or in a hybrid cloud spanning an on-premises implementation.
For how it’s only in one AWS Region, US West, and you can only buy by the hour. One-and-three-year subscriptions are due soon, as is a global rollout that will start soon and continue deep into 2018. There’s just one server type, too, and while vSphere lets you slice and dice that as it would any other server, there’s no hint of the varied instance types AWS and other clouds offer.
At least the server is pleasingly grunty. Each host has a pair of CPUs, 36 cores, 72 hyper-threads, 512GB of RAM, local flash storage (3.6TB cache, 10.7TB raw capacity tier). But you’ll need four hosts to build a cluster!
There is integration between VMware-on-AWS and some AWS services.
VMware will run and support the service, in contrast to the arrangement it has with IBM and the other ~4,300 vCloud Air Network partners that run vSphere-based clouds. Those partners get a new version of vCloud Director, plus more hardware partners ready to sell them servers ready to roll with Cloud Foundation. And perhaps some worry beads, for stress relief and/or prayer as VMware challenges them like never before, because the new service integrates with some AWS services. We’re told that the VMware service lives in the same data centres as services like Lambda, so piping them into apps will be low-latency.
In the past VMware partners have told El Reg they feel VMware’s cloud partnerships aren’t bad for business, because they get users talking about vSphere-powered clouds. Now we hear some are re-thinking that position, but the pricing for VMware on Amazon may well crimp their concerns, because it isn’t super-cheap.
Here’s the pricing scheme.
Remember: you’ll probably need at least four hosts, so actual costs will be rather more than the single-host cost.
VMware justifies these prices by saying they stack up well when compared to the total cost of ownership compared to either on-prem or public clouds.
Here’s the company’s math.
That calculation excludes bandwidth and IP address charges, and assumes VMs have a pair of vCPUs, 8GB RAM and 150GB of storage.
VMware’s attempt to build a public cloud failed, as did its early SaaS forays.
The company’s now turned that around, because the AWS deal gives it unrivalled scale, with perhaps-unsettling price.
The new SaaS offerings do two things:
VMware has done astoundingly well to keep Hyper-V’s market share small. But anyone who needs new servers or storage now has to consider either hyperconverged infrastructure or Azure Stack because both offer strong alternatives to traditional infrastructure. Azure Stack also makes hypervisors irrelevant and therefore also makes the idea of Windows-on-VMware look a bit archaic.
Starting with last week’s earnings call and already in pre-VMworld briefings, VMware’s counter argument is that it’s happy for you to use Azure in any form. So long as you don’t needlessly rip and replace perfectly good vSphere in order to buy in to Microsoft’s hybrid vision.
The new SaaS tools give you reasons not to ditch vSphere, by making multi-cloud wrangling easier and making vCenter the place you’ll do it. AppDefense helps, too, because it looks a useful tool that won’t hurt even if only deployed as one layer of a defense-in-depth strategy. It needs vCenter, too. And if vCenter is the place to do some security, and do multi-cloud management, it’s a lot harder to contemplate ejecting it. That the VMware/AWS tie-up has quickly gone beyond IaaS and into AWS’ services also suggests Virtzilla has found its way into a position of cloudy strength.
For now, anyway. Clouds move fast, and so do strategies to catch them.
Sponsored: The Joy and Pain of Buying IT – Have Your Say
Here is the original post:
VMware-on-AWS is live, and Virtzilla is now a proper SaaS player – The Register
Socionext Partners with Advantech to Offer High-Density, Low-Cost … – Design and Reuse (press release)
Scalable, Robust, Low-power, and Easily Deployable Solutions for service providers and other video intensive applications
SUNNYVALE, Calif. and MILPITAS, Calif., Aug. 31, 2017 — Socionext Inc., a world leader in hardware HEVC encoding, and Advantech, creator of innovative video acceleration solutions, today announced a strategic partnership to provide live hardware transcoding solutions for the data center supporting MPEG2, AVC (H.264), and HEVC (H.265).
Socionext real time dense transcode solution, also known as the “Media Cloud”, enables advanced HEVC compression technology and real-time transcoding capabilities for OTT applications. Socionext’s extended partnership with Advantech includes the integration of Socionext’s Media Cloud technology into Advantech’s VEGA 7000 Family of High Density Video Servers to enable agile and cost-effective live UHD cloud services for the new video-centric era.
“We are seeing an increasing need to lower the cost of ownership by media, telecom and internet companies that are seeking to address the ever-increasing mass consumption of streaming high-quality video,” said David Lin, VP of Video Solutions at Advantech. “Socionext, as our valued partner, is able to solve the power, density, and performance technical design requirements we are looking for in order for us to develop a cost-competitive, highly-efficient transcoding solution with adaptive bitrate (ABR) streaming capabilities for live cloud media service providers.”
The Advantech VEGA 7000 is a family of accelerated video processing servers which combine best video and IT practices within an off-the-shelf platform that has been optimized to efficiently scale throughput of high-density transcoding applications in live OTT and cloud workflows. Up to four VEGA-3318 accelerators can be integrated into a 1U server to deliver up to 32 x 4Kp60 live HEVC profiles per rack unit the highest density available in the market. This allows for large scale, energy and cost-efficient data center deployments that benefit from a 20X rack space and power reduction when compared to non-accelerated solutions. Advantech VEGA solutions for the data center minimize development efforts by providing a comprehensive software package that features Linux and Windows SDKs, an FFmpeg plug-in and virtualization-friendly drivers supporting OpenStack. Advantech also offers hardware and software design and customization services for maximum deployment flexibility.
“Advantech offers decades of expertise in complex hardware and software system integration and design services,” said Yasuhiro Wakimoto, VP of the Enterprise Solution Business Unit at Socionext. “Socionext and Advantech have a long history together providing solutions for “Live” transcode broadcasting and processing large volume of media data for video systems. This partnership further extends the close relationship.”
Advantech will demonstrate their VEGA 7000 Series of High Density Video Server for the Media Cloud at IBC 2017 in Hall 11, Booth C32, taking place at RAI, Amsterdam, from September 15-19, 2017. For more information, visit http://www.advantech.com/nc/spotlight/IBC2017 or email firstname.lastname@example.org.
Founded in 1983, Advantech is a leader in providing trusted, innovative products, services, and solutions. Advantech VEGA Video Platforms and PCIe Adapters are designed to boost video infrastructure performance from acquisition to distribution at the lowest power budget while fully complying with the media industry needs. By providing access to the latest 4K/8K UHD video processing and IP media technologies on commercial-off-the-shelf IT platforms we accelerate the deployment of next-generation, open and more efficient video solutions across a wide range of applications from broadcast encoding and high-density OTT transcoding to cloud, mobile and 360-degree video. Advantech’s standard portfolio can be tailored to meet a range of system requirements, significantly reducing time-to-market effort for our customers. For more information, visit http://www.video-acceleration.com.
About Socionext Inc.
Socionext is a new, innovative enterprise that designs, develops and delivers System-on-Chip products to customers worldwide. The company is focused on imaging, networking, computing and other dynamic technologies that drive today’s leading-edge applications. Socionext combines world-class expertise, experience, and an extensive IP portfolio to provide exceptional solutions and ensure a better quality of experience for customers. Founded in 2015, Socionext Inc. is headquartered in Yokohama, and has offices in Japan, Asia, United States and Europe to lead its product development and sales activities.
By Pete Eichorn
Although state and federal agencies are migrating steadily to the cloud, municipal governments have been slower to jump on the trend. Cloud technology presents a challenge for municipal agencies, which often dont know where to start, whom to ask for help, whether they need a third-party partner or whether laws and statutes even allow them to move data to the cloud.
Try before you buyFortunately, city and county governments dont have to go all in from the beginning. The cloud is a perfect small-scale testing ground. You can try cloud-based software as a service, or SaaS, which allows your team to use an existing software, delivered via a Web browser, and pay for it on a usage-time basis or with a monthly fee that grants access to designated staff members.
For example, a county that wants to replace its legacy financial system to better adhere to its states financial reporting requirements and eliminate its long-standing manual processes might test SaaS through the cloud. The county would gain access to an existing, third-party software product that could house its data, provide real-time access to financial transactions and reports and deliver other services the municipality might not even have anticipated.
The county would have avoided a large capital outlay. The cloud-based software would automatically scale when the county expected a larger demand on its financial services software and, if the software didnt meet the countys needs during the test period, the county would be free to walk away and try another solution.
The security questionData is no less secure in the cloud than on an internal server. That said, government cant outsource responsibility for its data. A cloud provider wont be responsible if there is a breach, so you must know what you are buying, where your data is and how it is being safeguarded. If you decide to test a cloud-based service, get the provider to disclose and include in your contract its terms and conditions, security limits, liabilities and responsibilities.
Your responsibilities to secure data still exist in the cloud as they do in traditional technologies. You must still comply with applicable laws and follow best practices, such as ensuring that user accounts and passwords are managed properly.
Which cloud?There are several kinds of clouds, and reasons you might choose one over another. Private clouds, maintained on-site, offer flexibility and control, and often are better suited when you want the clouds benefits and are willing to manage technology capabilities yourself. Public clouds, maintained via the internet on a cloud providers servers, are a good fit for accessing the latest technologies and for changing your financial management from a capital expense to an operating expense model.
Another option is hybrid cloud, which offers access to the best aspects of both public and private clouds. A hybrid cloud solution works well when you need to balance the needs of legacy systems and new, cloud capabilities.
Now is the timeMore and more, software providers are moving toward offering their products and services via the cloud. Innovation will focus on SaaS, and updates and support for legacy systems will become rarer.
So now is unquestionably the time for municipal governments to commit to understanding the potential benefits and dipping their toes in the water with test projects.
Research the benefits and individual services various cloud providers offer and determine which meet your needs. If you are short on internal resources, look for a government services provider to help you navigate the options and set out a cloud adoption plan that integrates with your existing systems, security models and goals.
To get connected and stay up-to-date with similar content from American City & County:Like us on FacebookFollow us on TwitterWatch us on YouTube
IBM hopes to raise its competitive profile in cloud services when it introduces new hardware and cloud infrastructure by the end of this year or early 2018.
The company will add a new collection of hardware and software products that deliver artificial intelligence (AI) and cloud-based services faster and more efficiently.
Among the server-based hardware technologies are 3D Torus, an interconnection topology for message-passing multicomputer systems, and new accelerators from Nvidia, along with advanced graphics processing unit (GPU) chips. Also included is Single Large Expensive Disk technology, a traditional disk technology currently used in mainframes and all-flash-based storage, according to sources familiar with the company’s plans.
The architecture achieves sub-20-millisecond performance latencies by eliminating routers and switches, and it embeds those capabilities into chips that communicate more directly with each other, one source said.
The new collection of hardware applies some of the same concepts as IBM’s Blue Gene supercomputer, which were among those used to create Watson. In the model of those special-purpose machines, the new system is designed specifically to do one thing: Deliver AI-flavored cloud-based services.
These technologies, which can work with both IBM Power and Intel chips in the same box, will be used only in servers housed in IBM’s data centers. IBM will not sell servers containing these technologies commercially to corporate users. The new technologies could reach IBM’s 56 data centers late this year or early next year.
IBM’s cloud business has grown steadily from its small base over the past three to four years to revenues of $3.9 billion in the company’s second quarter reported last month and $15.1 billion over the past 12 months. The company’s annual run rate for as-a-service revenues rose 32% from a year ago to $8.8 billion.
At the same time, sales of the company’s portfolio of cognitive solutions, with Watson at its core, took a step back, falling 1% in the second quarter after 3% growth in this year’s first quarter.
That doesn’t represent a critical setback, but it has caused some concern, because the company hangs much of its future growth on Watson.
Three years ago, IBM sunk $1 billion to set up its Watson business unit in the New York City borough Manhattan. IBM CEO Ginni Rometty has often cited lofty goals for the unit when claiming Watson would reach 1 billion consumers by the end of 2017, $1 billion in revenues by the end of 2018 and, eventually, $10 billion in revenue by an unnamed date. For IBM to achieve those goals, it requires a steady infusion of AI and machine learning technologies.
IBM executives remain confident, given the technical advancements in AI and machine learning capabilities built into Watson and a strict focus on corporate business users, while competitors — most notably Amazon — pursue consumer markets.
“All of our efforts around cognitive computing and AI are aimed at businesses,” said John Considine, general manager of cloud infrastructure at IBM. “This is why we have made such heavy investments in GPUs, bare-metal servers and infrastructure, so we can deliver these services with the performance levels corporate users will require.”
However, not everyone is convinced that IBM can reach its goals for cognitive cloud-based services, at least in the predicted time frames. And it will still be an uphill climb for Big Blue, as it looks to vie with cloud competitors faster out of the gate.
Lydia Leong, an analyst with Gartner, could not confirm details of IBM’s upcoming new hardware for cloud services, but pointed to the company’s efforts around a new cloud-oriented architecture dubbed Next Generation Infrastructure. NGI will be a new platform run inside SoftLayer facilities, but it’s built from scratch by a different team within IBM, she said.
My expectation is IBM will not have a long-term speed advantage with this — I’m not even sure they will have a short-term one. Lydia Leonganalyst, Gartner
IBM intends to catch up to the modern world of infrastructure with hardware and software more like those from competitors Amazon Web Services and Microsoft Azure, and thus deliver more compelling cloud-based services. NGI will be the foundation on which to build new infrastructure-as-a-service (IaaS) offerings, while IBM Bluemix, which remains a separate entity, will continue to run on top of bare metal.
Leong said she is skeptical, however, that any new server hardware will give the company a performance advantage to deliver cloud services.
“My expectation is IBM will not have a long-term speed advantage with this — I’m not even sure they will have a short-term one,” Leong said. “Other cloud competitors are intensely innovative and have access to the same set of technologies and tactical ideas, and they will move quickly.”
IBM has stumbled repeatedly with engineering execution in its cloud portfolio, which includes last year’s launch and demise of a new IaaS offering, OpenStack for Bluemix. “[IBM has] talked to users about this [NGI] for a while, but the engineering schedule keeps getting pushed back,” she said.
IBM now enters the cloud infrastructure market extremely late — and at a time when the core infrastructure war has been mostly won, Leong said. She suggested IBM might be better served to avoid direct competition with market leaders and focus its efforts where it has an established advantage and can differentiate with things like Watson.
Originally posted here:
IBM cooks up a hardware architecture for tastier cloud-based services – TechTarget
Due to improved analytics and the rapidly growing benefits of cognitive computing, data is becoming an even more critical growth asset central to all business operations.
Not only this, but with the increase of cyber attacks and new regulation such as General Data Protection Regulation (GDPR) soon coming into force, control over data is now at the top of all organisations agendas.
Now is the time to rethink architectures to enable companies to use data to drive digital transformation, while keeping full control of one of their key assets data.
In order to compete in a market being transformed by the availability of cloud services, many businesses are realising that they need complete control and transparency of their data management where it is, who has access to it and how it is protected.
>See also:Is business data AI compatible?
The way they collect, organise and activate this data will be essential to their future. This is why many businesses are now putting their data in the cloud ensuring their most critical asset is always in the hands of experts, with increased access to cognitive analytics.
Data on the moveIn todays digital environment, data is crucial to the operation of businesses. They need secure, compliant access to this through a multitude of channels. For example a bank needs access at the branch, online, and via mobile.
Wherever the operation is taking place, there needs to be a secure and compliant connection to the data centre, with consistent processes across the whole environment from mobile to mainframe.
Making this data available across geographies and channels only makes security more complex, with a growing range of endpoints and regulation to consider.
Data residency is important for organisations to consider when moving to the cloud. IBM through its growing cloud footprint in Europe, the UK and other global centres can offer clients the choice of where their data is stored.
>See also:Outlook increasingly cloudy: the mass migration of business data
The giant advocates for client choice of data residency while opposing government-mandated data localisation, with the exception of sensitive data for example related to national security or critical infrastructure.
The General Data Protection Regulation (GDPR) is another key piece of regulation to consider, which comes into force next year.
The aim of GDPR is to protect all EU citizens from privacy and data breaches in an increasingly data-driven world, massively increasing the scope of data protection and residency from the 1995 directive.
One of the major changes set to impact businesses is that serious breaches, which are likely to result in a risk for the rights and freedoms of individuals, must be reported within 72 hours. Those in breach can be fined up to 4% of annual global turnover, or 20 Million whichever is greater.
The current models in place for handling data will need to change and businesses are looking to sophisticated cloud providers to assist with this complexity.
With the prospect of significant fines for those who are in breach, security is top line priority for many organisations. While some still view cloud as less secure, enhanced security can be a key reason for moving to the cloud but not all clouds are created equally.
>See also:How can a business extract value from big data?
Some cloud providers can often offer more advanced security expertise than is available in house, managing end to end security needs from identity and access controls, to data protection and even threat intelligence.
This gives the end customer complete control of the cloud including the application and the hypervisor, all the way to the processor level.
Knowing who has access to what and where, otherwise known as identity and access management, is crucial.
Cloud identity management also known as Identity as a Service- gives businesses complete control over this, minimising the insider threat and ensuring there are no weak links in the chain.
Hardware can also prove vulnerable. With bare metal cloud servers, businesses can ensure that security and data geo-fencing is provided all the way to the processor level.
They get the flexibility and pay as you go benefits of cloud without sacrificing control of visibility over the environment.
With logical access and security controls, businesses can be assured that workloads are only being run on trusted hardware in the designated geography.
>See also:Using data analytics to improve business processes and reduce waste
Businesses can be further reassured that the cloud is safe and secure with the introduction of the European Cloud Code of Conduct.
The Code is rigorous companies that sign up must meet all of the requirements with no room for opt-outs or exceptions.
Not only does the Code align with the current legal framework for data protection in Europe, but it also aligns with core elements of the GDPR.
One of the industries most concerned with security is financial services. Even businesses in this highly regulated industry are now expanding their cloud solutions.
While there is still a demand for in-house IT to run certain core systems and store particular data types, hybrid cloud is increasingly being seen as the most effective solution for this type of industry, integrating local, public and shared options.
>See also:The value of data driving business innovation and acceleration
One bank already taking advantage of these benefits is Boursorama, a subsidiary of the Socit Gnrale Group and leader in online banking, online brokerage and financial information services.
By deploying its services through a cloud infrastructure, Boursorama has been able to improve management of IT resources by quickly activating servers based on specific client needs and usage. It has also been able to strengthen the security and resilience of its websites.
Through a global, 247 managed backup solution for its back office that leverages one of IBMs business resiliency centres, Boursorama has been able to speed up recovery times by providing faster access to the crucial applications and data that are core to its business.
In a data led economy, businesses today cant afford to lose track of their most valuable asset. So the question is, do you know where your data is?
The UKs largest conference fortechleadership,TechLeadersSummit, returns on 14 September with 40+ top execs signed up to speak about the challenges and opportunities surrounding the most disruptive innovations facing the enterprise today.Secure your place at this prestigious summit byregisteringhere
See the original post:
Where does a business’s data live? – Information Age