|By Ted Alford, Gwen Morton||
|October 26, 2009 02:00 AM EDT||
Government Cloud Computing on Ulitzer
The President's budget for fiscal year 2010 (FY10) includes $75.8B in information technology (IT) spending, which is a 7-percent increase from FY09. Of this, at least $20B will be spent on IT infrastructure investments.  The FY11 budget for IT is projected to be nearly $88B. The government is actively seeking ways to reduce IT costs, and the FY10 budget request highlights opportunities for the federal government to achieve significant long-term cost savings through the adoption of cloud computing technologies:
"Of the investments that will involve up-front costs to be recouped in outyear savings, cloud-computing is a prime case in point. The Federal Government will transform its Information Technology Infrastructure by virtualizing data centers, consolidating data centers and operations, and ultimately adopting a cloud-computing business model. Initial pilots conducted in collaboration with Federal agencies will serve as test beds to demonstrate capabilities, including appropriate security and privacy protection at or exceeding current best practices, developing standards, gathering data, and benchmarking costs and performance. The pilots will evolve into migrations of major agency capabilities from agency computing platforms to base agency IT processes and data in the cloud. Expected savings in the outyears, as more agencies reduce their costs of hosting systems in their own data centers, should be many times the original investment in this area." 
The language in the budget makes three key points: (1) up-front investment will be made in cloud computing, (2) long-term savings are expected, and (3) the savings are expected to be significantly greater than the investment costs.
Booz Allen Hamilton has created a detailed cost model that can create life-cycle cost (LCC) estimates of public, private, and hybrid clouds. We used this model, and our extensive experience in economic analysis of IT programs, to arrive at a first-order estimate of each of the three key points in the President's budget. Overall, it appears likely that the expectations highlighted in the budget can be met, but several factors could affect the overall degree of economic benefit.
The government's adoption of this new IT model warrants careful consideration of the model's broad economic implications-including the potential long-term benefits in terms of cost savings and avoidance as well as the near-term costs and other impacts of a transition from the current environment. Factors such as the number and rate of federal agencies adopting cloud computing, the length of their transitions to cloud computing, and the cloud computing deployment model (public, private, or hybrid) all will affect the total costs, potential benefits, and time required for the expected benefits to offset the investment costs.
Booz Allen developed a first-order economic analysis by considering how agencies might migrate to a cloud-based environment and what the costs and potential savings might be under a variety of scenarios. Specifically, given long-standing efforts to protect the privacy and security of the federal government's data and systems, a key variable will be whether agencies take advantage of public clouds, build their own private clouds, or adopt a hybrid approach., The focus was on Cloud Computing infrastructure services as these tend to represent a relatively more consistent set of costs/investments/operating requirements across all agencies. We made some high-level, simplifying assumptions in our initial analysis:
- There is an existing data center(s) currently operational that is a baseline for economic comparison of migrating to a cloud environment.
- Existing application software will migrate with the infrastructure to the cloud. Application software support costs remain out of scope.
- Migration decisions will be made at the department or agency (rather than bureau) level in order to aggregate demand and drive scale efficiencies.
- We assume the perceived sensitivity of an agency's mission and data will be a primary factor (though by no means the only factor) driving its decisions on which path to follow..
Next, we developed three high-level scenarios that represent potential migration paths. The three scenarios are as follows:
Scenario 1: Public Cloud Adopters
Key Agency Characteristic: Migrates low-sensitivity data to an existing public cloud.
Assumptions: Transition to the new cloud environment will occur steadily over 3 years; workload remains constant (i.e., no increase in capacity demand).
Scenario 2: Hybrid Cloud Adopters
Key Agency Characteristic: Uses a private cloud solution to handle the majority of its IT workload; also uses a public cloud solution to provide "surge" support and/or support for low-sensitivity data.
Assumptions: Seventy-five percent of the IT server workload will migrate to a private cloud, and the remaining 25 percent will transition to a public cloud; transition to the new cloud environments will occur steadily over 3 years; existing facilities will be used (i.e., no new investment is required in physical facilities); workload remains constant (i.e., no increase in capacity demand).
Scenario 3: Private Cloud Adopters
Key Agency Characteristic: Builds its own private cloud solution or participates in an interagency cloud solution (i.e., community cloud). Broad mission sensitivity results in the need to maintain control of infrastructure and data.
Assumptions: Transition to the new cloud environment will occur steadily over 3 years; existing facilities will be used (i.e., no new investment is required in physical facilities); workload remains constant (i.e., no increase in capacity demand).
Agencies publicly report only their "consolidated" IT infrastructure expenditures, which include end-user support systems (e.g., desktops, laptops) and telecommunications. Additional spending on application-specific IT infrastructure is typically rolled up into individual IT investments. In an effort to isolate data center costs, we extrapolated findings based on our experience with actual federal data centers. Specifically, we developed a "representative" agency data center profile that serves as a useful proxy for other agencies and enables us to explore the potential savings of a migration to cloud computing under the scenarios described above. Although agencies of similar size can have very different IT infrastructure profiles, we modeled an agency with a classic standards-based web application infrastructure. For our representative agency, we began with an assumption that a Status Quo (SQ) data center containing 1,000 servers with no virtualization is already operational.  The results at different scales are shown in our analysis.
Using a Booz Allen proprietary cloud computing cost and economic model that employs data collected internally, data from industry, and parametric estimating techniques, we estimated the LCCs for our representative agency to migrate its IT infrastructure (i.e., its server hardware and software) to the cloud under each of the three scenarios described above. We compared these costs to the LCCs of the SQ scenario (i.e., no cloud migration).  We also calculated three common metrics to analyze each scenario's potential economic benefits. These metrics allowed us to evaluate the three elements of the business case in the President's budget and estimate the absolute and relative benefits, as well as the time over which the outyear savings will pay back the investment costs.
The three key metrics in our analysis are as follows:
- Net Present Value (NPV) is calculated as each cloud scenario's discounted net benefits (i.e., the cloud scenario's reduced operations and support [O&S] costs relative to the SQ environment's O&S costs) minus the cloud's discounted one-time investment costs. A positive dollar figure indicates a positive economic benefit versus the SQ environment. NPV is an absolute economic metric.
- Benefit-to-Cost Ratio (BCR) is calculated as each cloud scenario's discounted net benefits divided by its discounted investment costs. A number greater than 1.0 indicates a positive economic benefit versus the SQ environment. BCR is a relative economic metric.
- Discounted Payback Period (DPP) reflects the number of years (from FY10) it takes for each scenario's accumulated annual benefits to equal its total investment costs.
The top portion of Exhibit 1 shows the analysis results. This exhibit presents the one-time investment phase costs as well as the recurring O&S phase costs for each scenario with a 13-year life cycle (3-year investment phase and 10-year steady-state O&S phase) from FY10 through FY22.
Assuming a 3-year transition period for each scenario, investment costs are expected to be incurred from FY10 to FY12 and include (depending on the scenario) hardware procurement and commercial off-the-shelf (COTS) software license fees; contractor labor required for installation, configuration, and testing; and technical and planning support (i.e., system engineering and program management costs) before and during the cloud migration. Because the SQ reflects an operational steady state, no investment costs are estimated for that scenario. Although the public cloud scenario does not present any up-front investment costs for hardware or software procurement, it does require program planning and technical support, support for porting applications over to the new cloud environment, and testing support to ensure programs and applications are working correctly in the new environment.
Recurring O&S costs "ramp up" for all cloud scenarios beginning in FY10 and enter steady state in FY13, continuing through FY22. For private clouds, these costs include hardware and software maintenance, periodic replacement/license renewal costs, system operations labor support costs, and IT power and cooling costs. For hybrid clouds, the O&S costs include the same items as the private cloud (albeit on a reduced scale), as well as the unit consumption costs of IT services procured from the public cloud. For public cloud scenarios, the O&S costs are the unit costs of services procured from the cloud provider and a small amount of IT support labor for the cloud provider to communicate any service changes or problems. In all three cloud scenarios, a significant portion of the O&S costs are incurred while phasing out the SQ environment during the transition. The SQ phase-out costs "ramp down" from FY10 to FY12, dove-tailing with the ramp up of the new clouds' O&S costs. Not surprisingly, the total LCCs are lowest for the public cloud scenario and highest for the private cloud scenario, with the hybrid cloud scenario's LCCs falling in the middle.
The economic analysis confirms that the projected NPV and BCR for all three scenarios are significant relative to the SQ environment. Once the cloud migrations are completed, our model estimates annual O&S savings in the 65-85 percent range, with the lower end corresponding to the private cloud scenario and the upper end corresponding to the public cloud scenario. These percentages can be applied to overall federal IT spending for data centers to estimate the potential absolute savings across the federal government. (As part of the Information Technology Infrastructure Line of Business [ITI LoB] initiative, General Services Administration [GSA] is coordinating a benchmarking effort across the government. If those figures are made public, a total dollar savings estimate will be possible).
Our model shows that the net benefits and payback periods for agencies adopting the hybrid cloud scenario are closer to those for the private cloud than the public cloud. This variation is largely a result of our assumption that 75 percent of the current server workload would migrate to a private cloud and only 25 percent would transition to the public cloud. If we were instead to assume the opposite mix (i.e., 25 percent of the workload migrating to a private cloud and 75 percent to a public cloud), the hybrid scenario economic results would be closer to the public cloud results.
We conducted a sensitivity analysis on several of the variables in our cost model to determine the major drivers for cloud economics. The two most influential factors driving the economic benefits are (1) the reduction in hardware as a smaller number of virtualized servers in the cloud replace physical servers in the SQ data center and (2) the length of the cloud migration schedule. Exhibits 2, 3, and 4 show the results of varying these factors.
In practice, several factors could cause agencies to realize lower economic benefits than our estimates suggest. One factor is underestimation of the costs associated with the investment or O&S phase for the cloud scenarios. Another factor is server utilization rates (both in the current environment and the new cloud environment). Our analysis assumes an average utilization rate of 12 percent of available CPU capacity in the SQ environment and 60 percent in the virtualized cloud scenarios. This difference in server utilization, in turn, enables a large reduction in the number of servers (and their associated support costs) required in a cloud environment to process the same workload relative to the SQ environment. Agencies with server utilization rates that are already relatively high should expect lower potential savings from a virtualized cloud environment.
The charts indicate two key takeaways:
- Scale is important: The economic benefit increases as virtualized servers replace larger numbers of underutilized servers.
- Time is money: Because of the cost of parallel IT operations (i.e., cloud and non-cloud), the shorter the server migration schedule, the greater the economic benefits.
These findings, in turn, lead us to the following recommendations for agencies and policymakers contemplating a cloud migration:
- It is more cost-effective to group smaller existing data centers together into as large a cloud as possible, rather than creating several smaller clouds.
To reduce the cost of running parallel operations, organization should properly plan for and then migrate to the new cloud environment as quickly as possible. The three lines in Exhibit 5 show (in this case, for the public cloud) that the BCR goes down rapidly and the DPP increases as the transition time increases.
A few agencies are already moving quickly to explore cloud computing solutions and are even redirecting existing funds to begin implementations. However, for most of the federal government, the timeframe for redirecting IT funding to support cloud migrations is likely to be at least 1-2 years, given that agencies formulate budgets 18 months before receiving appropriations.
Specifically, an agency develops IT investment requests each spring and submits them to OMB in September, along with the agency's program budget request, for the following government fiscal year. OMB reviews agency submissions in the fall and can implement funding changes via passback decisions (generally in late November) before submitting the President's budget to the Congress in February. Theoretically, the earliest opportunity for OMB to push agencies to revise their IT budgets to support a transition to the cloud will be fall 2009; however, agencies typically only have about 1 month to incorporate changes to their IT portfolios during passback. To give GSA and OMB time to develop more detailed guidance, as well as necessary procurement mechanisms and vehicles, it is more likely that OMB will direct or encourage agencies to plan for cloud migrations during the FY12 budget cycle (starting in the spring of 2010).
Other Considerations with Potential Economics Effects
When deciding on moving to the cloud, agencies need to consider some additional technical aspects of cloud computing and their potential impact on their organization. Such areas include but are not limited to data security, software migration, technical architectures, and the skill set of the IT workforce.
All government organizations struggle with ensuring that the data they have remains secure and adheres to current policies and regulations. Because data security is such a critical issue, cloud providers will be required to address it in their products and services, and should be able to tailor the level of security to meet demand. Additionally, by centralizing data and servers, a cloud environment will allow for easier detection and investigation of incidents, and allow enabling IT staff to replicate and address them efficiently.
However, there are currently no security standards for cloud computing, and until such standards have been developed, and used effectively to measure provider services and enforce accountability, any failures will fall on the agency's in-house IT organization. In awareness of this reality, organizations should be careful about putting mission-critical and core processes into a public cloud, and private cloud architectures should be designed to minimize any security concerns while realizing the benefits of cloud optimization.
Service Oriented Architecture
As the government moves towards embracing Service-Oriented Architecture (SOA), cloud computing will optimize the benefits of those investments. Cloud computing is inherently a Service Oriented Architecture and implementing the private clouds will provide for more control over data, security and privacy.
Migration of Applications to the Cloud
This article identifies the financial benefits of migrating the IT infrastructure to the cloud.
Cloud architectures and service delivery models will lead to changing needs for technical skills amongst agencies' IT workforces. CIOs will need to plan to conduct or refresh workforce assessments and training, as well as set aside the necessary funding, to ensure technical staff are trained on cloud architecture, implementation and operations.
Economic Influence on Policy
From an economic perspective, GSA and OMB can take a number of steps to maximize the probability that the cloud computing business model can work in the federal government; i.e., that it can achieve its objective of enabling significant cost savings. These steps promote information sharing and transparency in the realistic costs and benefits of various cloud models, as well as establishing the necessary policy and contracting frameworks. Because scale is a key variable affecting both costs and benefits, policy guidance regarding scale considerations will be particularly critical (e.g., determining how much flexibility, if any, agencies and departments have to create private clouds at the bureau and/or interagency level).
As a cloud "storefront," GSA should conduct due diligence reviews to establish that public cloud providers, once identified, indeed offer highly efficient, highly scalable (both up and down) usage-based pricing beyond traditional managed services (e.g., by comparing proposed rates against commercial benchmarks). GSA should also work with potential providers to ensure agencies can readily understand service definitions, service levels, terms, conditions, and pricing. These steps will provide transparency to facilitate agencies' ability to compare potential provider pricing against their legacy operations costs-an essential component of building a credible business case for any type of cloud migration. In earlier shared services initiatives, such as financial management, the lack of such standardized information on pricing and service levels in the first few years proved a major impediment to progress, as agencies faced decisions about alternative solutions that were often based on unreliable cost data from potential vendors.
Finally, GSA will need to establish and communicate its own schedule for cloud services founded in the pricing for the services with different cloud venders..
Summary of Key Observations
Although cloud computing offers potentially significant savings to federal agencies by reducing their expenditures on server hardware and associated support costs, chief information officers, policymakers, and other interested parties should bear in mind a number of practical considerations:
- It will take, on average, 18-24 months for most agencies to redirect funding to support this transition, given the budget process.
- Some up-front investment will be required, even for agencies seeking to take advantage of public cloud options.
- Implementations may take several years, depending on the size of the agency and the complexity of the cloud model it selects (i.e., public, private, or hybrid).
- It could take as long as 4 years for the accumulated savings from agency investments in cloud computing to offset the initial investment costs; this timeframe could be longer if implementations are improperly planned or inefficiently executed.
Given these observations, we offer the following recommendations:
- OMB, GSA, and other organizations, such as National Institute of Standards and Technology (NIST), should provide timely, well-coordinated support-in the form of necessary standards, guidance, policy decisions, and issue resolution-to ensure agencies have the necessary tools to efficiently plan and carry out migrations to cloud environments. As the length of the migration period increases, the potential economic benefits of the migration decrease.
- OMB and GSA should seek to identify those agencies with the highest near-term IT costs and expedite their migration to the cloud.
- To encourage steady progress, OMB should establish a combination of incentives and disincentives; e.g., consider allowing agencies to retain a small percentage of any savings realized from cloud computing for investments in future initiatives. To monitor progress and heighten transparency and accountability, OMB could incorporate cloud-related metrics into the new government-wide IT dashboard.
- Agencies should consider which of the high-level scenarios described in this article best suits their needs, with the understanding that regardless of scenario chosen, proper planning and efficient execution are critical success factors from an economic perspective.
- Given the significant impact of scale efficiencies, agencies selecting a private cloud approach should fully explore the potential for interdepartmental and interagency collaboration and investment (consistent with emerging OMB and GSA guidance). This, in effect, leads to the fourth cloud deployment model-the community cloud. A community cloud is a collaboration between private cloud operators to share resources and services.
- Agencies should identify the aspects of their current IT workload that can be transitioned to the cloud in the near term to yield "early wins" to help build momentum and support for the migration to cloud computing.
Cloud computing has received executive backing and offers clear opportunities for agencies to significantly reduce their growing data center and IT hardware expenditures. However, for the government to achieve the envisioned savings, organizations charged with oversight, such as OMB, NIST, and GSA, will have to facilitate progress, and departments and agencies will have to carefully select and plan for future cloud scenarios that yield the best tradeoffs among their respective costs, benefits, and risks.
- Figures from INPUT data for the FY10 President's budget; of the $20B in expenditures categorized as office automation and IT infrastructure spending, about $12.2B is spent on major IT investments, with the remainder on non-majors. Additional expenditures on application-specific IT infrastructure are typically reported as part of individual IT investments.
- President's budget, FY10 (Analytical Perspectives).
- The 1,000 servers are broken down in our cost model by server processing capacity (small, medium, and large) based on proportions consistent with our experience.
- Our model focuses on the costs that a cloud migration will most likely directly affect; i.e., costs for server hardware (and associated support hardware, such as internal routers and switches, rack hardware, cabling, etc.), basic server software (OS software, standard backup management, and security software), associated contractor labor for engineering and planning support during the transition phase, hardware and software maintenance, IT operations labor, and IT power/cooling costs. It does not address other costs that would be less likely to vary significantly between cloud scenarios, such as storage, application software, telecommunications, or WAN/LAN. In addition, it does not include costs for government staff. Further, for simplicity we removed facilities cost from the analysis
|jhbeil 10/21/09 03:51:00 PM EDT|
so when is "cloudonomics" going to hit the bookshelves?
|Phillip Hallam-Baker 10/20/09 09:30:00 PM EDT|
Looking at the numbers in the article a little further, it is assumed that the utilization rate will increase from 16% to 60% and that the reduction in the number of machines is the reason for the purported 60% cost saving.
The only way I can make those numbers work is if it is assumed that 80% of the costs in a data center are driven by nothing more than the number of machines in the data center that are powered.
This seems to be an absurdly high assumption to me.
|Phillip Hallam-Baker 10/19/09 05:07:00 PM EDT|
I found the basic assumptions in this article to be unsupported. It is really easy to assume 65% savings from an infrastructure change if you ignore most of the costs of making the change.
I examine this in more detail on my blog.
I think this type of article will do great damage to cloud computing as it sets out claims that are simply ludicrous and will not be believed. It is entirely credible that newly deployed software services will be cheaper when designed for cloud deployment. It is not credible that anyone should expect to save a single dollar by taking a deployed application that does not otherwise need changing and throwing it into the cloud.
Once hardware costs are sunk, they are sunk. thus there are no savings to be won through 'migration' if you are a large corporation or a government agency. There will be real savings, but they will be modest and come gradually.
The savings from cloud computing will be for the smaller enterprise right down to the small business which does not even have a machine room let alone a data center. There the savings are real and dramatic. But let's not get cloud computing dismissed as hype with unsupported claims.
Ask someone to architect an Internet of Things (IoT) solution and you are guaranteed to see a reference to the cloud. This would lead you to believe that IoT requires the cloud to exist. However, there are many IoT use cases where the cloud is not feasible or desirable. In his session at @ThingsExpo, Dave McCarthy, Director of Products at Bsquare Corporation, will discuss the strategies that exist to extend intelligence directly to IoT devices and sensors, freeing them from the constraints of ...
Jul. 1, 2016 09:12 AM EDT
The IoT is changing the way enterprises conduct business. In his session at @ThingsExpo, Eric Hoffman, Vice President at EastBanc Technologies, discussed how businesses can gain an edge over competitors by empowering consumers to take control through IoT. He cited examples such as a Washington, D.C.-based sports club that leveraged IoT and the cloud to develop a comprehensive booking system. He also highlighted how IoT can revitalize and restore outdated business models, making them profitable ...
Jul. 1, 2016 09:00 AM EDT Reads: 611
IoT offers a value of almost $4 trillion to the manufacturing industry through platforms that can improve margins, optimize operations & drive high performance work teams. By using IoT technologies as a foundation, manufacturing customers are integrating worker safety with manufacturing systems, driving deep collaboration and utilizing analytics to exponentially increased per-unit margins. However, as Benoit Lheureux, the VP for Research at Gartner points out, “IoT project implementers often ...
Jul. 1, 2016 08:45 AM EDT Reads: 759
When people aren’t talking about VMs and containers, they’re talking about serverless architecture. Serverless is about no maintenance. It means you are not worried about low-level infrastructural and operational details. An event-driven serverless platform is a great use case for IoT. In his session at @ThingsExpo, Animesh Singh, an STSM and Lead for IBM Cloud Platform and Infrastructure, will detail how to build a distributed serverless, polyglot, microservices framework using open source tec...
Jul. 1, 2016 08:30 AM EDT Reads: 750
The idea of comparing data in motion (at the sensor level) to data at rest (in a Big Data server warehouse) with predictive analytics in the cloud is very appealing to the industrial IoT sector. The problem Big Data vendors have, however, is access to that data in motion at the sensor location. In his session at @ThingsExpo, Scott Allen, CMO of FreeWave, discussed how as IoT is increasingly adopted by industrial markets, there is going to be an increased demand for sensor data from the outermos...
Jul. 1, 2016 08:00 AM EDT Reads: 522
CenturyLink has announced that application server solutions from GENBAND are now available as part of CenturyLink’s Networx contracts. The General Services Administration (GSA)’s Networx program includes the largest telecommunications contract vehicles ever awarded by the federal government. CenturyLink recently secured an extension through spring 2020 of its offerings available to federal government agencies via GSA’s Networx Universal and Enterprise contracts. GENBAND’s EXPERiUS™ Application...
Jul. 1, 2016 08:00 AM EDT Reads: 542
A strange thing is happening along the way to the Internet of Things, namely far too many devices to work with and manage. It has become clear that we'll need much higher efficiency user experiences that can allow us to more easily and scalably work with the thousands of devices that will soon be in each of our lives. Enter the conversational interface revolution, combining bots we can literally talk with, gesture to, and even direct with our thoughts, with embedded artificial intelligence, wh...
Jul. 1, 2016 07:15 AM EDT Reads: 1,298
"delaPlex is a software development company. We do team-based outsourcing development," explained Mark Rivers, COO and Co-founder of delaPlex Software, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 1, 2016 07:15 AM EDT Reads: 681
"We work in the area of Big Data analytics and Big Data analytics is a very crowded space - you have Hadoop, ETL, warehousing, visualization and there's a lot of effort trying to get these tools to talk to each other," explained Mukund Deshpande, head of the Analytics practice at Accelerite, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 1, 2016 12:30 AM EDT Reads: 734
Cloud Expo, Inc. has announced today that Andi Mann returns to 'DevOps at Cloud Expo 2016' as Conference Chair The @DevOpsSummit at Cloud Expo will take place on November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. "DevOps is set to be one of the most profound disruptions to hit IT in decades," said Andi Mann. "It is a natural extension of cloud computing, and I have seen both firsthand and in independent research the fantastic results DevOps delivers. So I am excited t...
Jul. 1, 2016 12:00 AM EDT Reads: 611
The cloud promises new levels of agility and cost-savings for Big Data, data warehousing and analytics. But it’s challenging to understand all the options – from IaaS and PaaS to newer services like HaaS (Hadoop as a Service) and BDaaS (Big Data as a Service). In her session at @BigDataExpo at @ThingsExpo, Hannah Smalltree, a director at Cazena, provided an educational overview of emerging “as-a-service” options for Big Data in the cloud. This is critical background for IT and data profession...
Jun. 30, 2016 04:00 PM EDT Reads: 544
Machine Learning helps make complex systems more efficient. By applying advanced Machine Learning techniques such as Cognitive Fingerprinting, wind project operators can utilize these tools to learn from collected data, detect regular patterns, and optimize their own operations. In his session at 18th Cloud Expo, Stuart Gillen, Director of Business Development at SparkCognition, discussed how research has demonstrated the value of Machine Learning in delivering next generation analytics to imp...
Jun. 30, 2016 04:00 PM EDT Reads: 1,041
Whether your IoT service is connecting cars, homes, appliances, wearable, cameras or other devices, one question hangs in the balance – how do you actually make money from this service? The ability to turn your IoT service into profit requires the ability to create a monetization strategy that is flexible, scalable and working for you in real-time. It must be a transparent, smoothly implemented strategy that all stakeholders – from customers to the board – will be able to understand and comprehe...
Jun. 30, 2016 02:45 PM EDT Reads: 418
Connected devices and the industrial internet are growing exponentially every year with Cisco expecting 50 billion devices to be in operation by 2020. In this period of growth, location-based insights are becoming invaluable to many businesses as they adopt new connected technologies. Knowing when and where these devices connect from is critical for a number of scenarios in supply chain management, disaster management, emergency response, M2M, location marketing and more. In his session at @Th...
Jun. 30, 2016 01:30 PM EDT Reads: 1,392
In his keynote at 18th Cloud Expo, Andrew Keys, Co-Founder of ConsenSys Enterprise, provided an overview of the evolution of the Internet and the Database and the future of their combination – the Blockchain. Andrew Keys is Co-Founder of ConsenSys Enterprise. He comes to ConsenSys Enterprise with capital markets, technology and entrepreneurial experience. Previously, he worked for UBS investment bank in equities analysis. Later, he was responsible for the creation and distribution of life sett...
Jun. 30, 2016 01:00 PM EDT Reads: 1,533
IoT is rapidly changing the way enterprises are using data to improve business decision-making. In order to derive business value, organizations must unlock insights from the data gathered and then act on these. In their session at @ThingsExpo, Eric Hoffman, Vice President at EastBanc Technologies, and Peter Shashkin, Head of Development Department at EastBanc Technologies, discussed how one organization leveraged IoT, cloud technology and data analysis to improve customer experiences and effi...
Jun. 30, 2016 11:30 AM EDT Reads: 677
Basho Technologies has announced the latest release of Basho Riak TS, version 1.3. Riak TS is an enterprise-grade NoSQL database optimized for Internet of Things (IoT). The open source version enables developers to download the software for free and use it in production as well as make contributions to the code and develop applications around Riak TS. Enhancements to Riak TS make it quick, easy and cost-effective to spin up an instance to test new ideas and build IoT applications. In addition to...
Jun. 30, 2016 11:15 AM EDT Reads: 763
The cloud market growth today is largely in public clouds. While there is a lot of spend in IT departments in virtualization, these aren’t yet translating into a true “cloud” experience within the enterprise. What is stopping the growth of the “private cloud” market? In his general session at 18th Cloud Expo, Nara Rajagopalan, CEO of Accelerite, explored the challenges in deploying, managing, and getting adoption for a private cloud within an enterprise. What are the key differences between wh...
Jun. 30, 2016 11:00 AM EDT Reads: 1,138
Presidio has received the 2015 EMC Partner Services Quality Award from EMC Corporation for achieving outstanding service excellence and customer satisfaction as measured by the EMC Partner Services Quality (PSQ) program. Presidio was also honored as the 2015 EMC Americas Marketing Excellence Partner of the Year and 2015 Mid-Market East Partner of the Year. The EMC PSQ program is a project-specific survey program designed for partners with Service Partner designations to solicit customer feedbac...
Jun. 30, 2016 10:45 AM EDT Reads: 727
There are several IoTs: the Industrial Internet, Consumer Wearables, Wearables and Healthcare, Supply Chains, and the movement toward Smart Grids, Cities, Regions, and Nations. There are competing communications standards every step of the way, a bewildering array of sensors and devices, and an entire world of competing data analytics platforms. To some this appears to be chaos. In this power panel at @ThingsExpo, moderated by Conference Chair Roger Strukhoff, Bradley Holt, Developer Advocate a...
Jun. 30, 2016 10:15 AM EDT Reads: 1,004