Archivo de la categoría: IoT

The IoT in Palo Alto: connecting America’s digital city

jonathan_reichental_headshot_banffPalo Alto is not your average city. Established by the founder of Stanford University, it was the soil from which Google, Facebook, Pinterest and PayPal (to name a few) have sprung forth. Indeed, Palo Alto has probably done more to transform human life in the last quarter century than any other. So, when we think of how the Internet of Things is going to affect life in the coming decades, we can be reasonably sure where much of expected disruption will originate.

All of which makes Palo Alto a great place to host the first IoT Data Analytics & Visualization event (February 9 – 11, 2016). Additionally fitting: the event is set to be kicked off by Dr. Jonathan Reichental, the city’s Chief Information Officer: Reichental is the man entrusted with the hefty task of ensuring the city is as digital, smart and technologically up-to-date as a place should be that has been called home by the likes of Steve Jobs, Mark Zuckberg, Larry Page and Sergey Brin.

Thus far, Reichental’s tenure has been a great success. In 2013, Palo Alto was credited with being the number one digital city in the US, and has made the top five year upon year – in fact, it so happens that, following our long and intriguing telephone interview, Reichental is looking forward to a small celebration to mark its latest nationwide ranking.

BCN: Jonathan, you’ve been Palo Alto’s CIO now for four years. What’s changed most during that time span?

Dr Jonathan Reichental: I think the first new area of substance would be open government. I recognise open government’s been a phenomenon for some time, but over the course of the last four years, it has become a mainstream topic that city and government data should be easily available to the people. That it should be machine readable, and that an API should be made available to anyone that wants the data. That we have a richer democracy by being open and available.

We’re still at the beginning however. I have heard that there are approximately 90,000 public agencies in the US alone. And every day and week I hear about a new federal agency or state or city of significance who are saying, ‘you can now go to our data portal and you can access freely the data of the city or the public agency. The shift is happening but it’s got some way to go.

Has this been a purely technical shift, or have attitudes had to evolve as well?

I think if you kind of look at something like cloud, cloud computing and cloud as a capability for government – back when I started ‘cloud’ was a dirty word. Many government leaders and government technology leaders just weren’t open to the option of putting major systems off-premise. That has begun to shift quite positively.

I was one of the first to say that cloud computing is a gift to government. Cloud eliminates the need to have all the maintenance that goes with keeping systems current and keeping them backed up and having disaster recovery. I’ve been a very strong proponent of that.

Then there’s social media  – government has fully embraced that now, having been reluctant early on. Mobile is beginning to emerge though it’s still very nascent. Here in Palo Alto we’re trying to make all services that make sense accessible via smart phone. I call it ‘city in a box.’ Basically, bringing up an app on the smart phone you should be able to interact with government – get a pet license, pay a parking fee, pay your electrical bill: everything should really be right there on the smartphone, you shouldn’t need to go to City Hall for many things any more.

The last thing I’d say is there has been an uptake in community participation in government. Part of it is it’s more accessible today, and part of it is there’s more ways to do so, but I think we’re beginning also to see the fruits of the millennial generation – the democratic shift in people wanting to have more of a voice and a say in their communities. We’re seeing much more in what is traditionally called civic engagement. But ‘much more’ is still not a lot. We need to have a revolution in this space for there to be significant change to the way cities operate and communities are effective.

Palo Alto is hosting the IoT Data Analytics & Visualization in February. How have you innovated in this area as a city?

One of the things we did with data is make it easily available. Now we’re seeing a community of people in the city and beyond, building solutions for communities. One example of that is a product called Civic Insight. This app consumes the permit data we make available and enables users to type in an address and find out what’s going on in their neighbourhood with regard to construction and related matters.

That’s a clear example of where we didn’t build the thing, we just made the data available and someone else built it. There’s an economic benefit to this. It creates jobs and innovation – we’ve seen that time and time again. We saw a company build a business around Palo Alto releasing our budget information. Today they are called OpenGov, and they sell the solution to over 500 cities in America, making it easy for communities to understand where their tax payer dollars are being spent. That was born and created in Palo Alto because of what we did making our data available.

Now we get to today, and the Internet of Things. We’re still – like a lot folks, especially in the government context – defining this. It can be as broad or as narrow as you want. There’s definitely a recognition that when infrastructure systems can begin to share data between each other, we can get better outcomes.

The Internet of Things is obviously quite an elastic concept, but are there areas you can point to where the IoT is already very much a reality in Palo Alto?

The clearest example I can give of that today is our traffic signal system here in the city. A year-and-a-half ago, we had a completely analogue system, not connected to anything other than a central computer, which would have created a schedule for the traffic signals. Today, we have a completely IP based traffic system, which means it’s basically a data network. So we have enormous new capability.

For example, we can have schedules that are very dynamic. When schools are being let out traffic systems are one way, at night they can be another way, you can have very granular information. Next you can start to have traffic signals communicate with each other. If there is a long strip of road and five traffic systems down there is some congestion, all the other traffic signals can dynamically change to try and make the flow better.

It goes even further than this. Now we can start to take that data – recording, for example, the frequency and volume of vehicles, as well as weather, and other ambient characteristics of the environment – and we can start to send this to the car companies. Here at Palo Alto, almost every car company has their innovation lab. Whether it’s Ford, General Motors, Volkswagen, BMW, Google (who are getting into the car business now) – they’re all here and they all want our data. They’re like: ‘this is interesting, give us an API, we’ll consume it into our data centres and then we’ll push into cars so maybe they can make better decisions.’

You have the Internet of Things, you’ve got traffic signals, cloud analytics solutions, APIs, and cars as computers and processors. We’re starting to connect all these related items in a way we’ve never done before. We’re going to follow the results.

What’s the overriding ambition would you say?

We’re on this journey to create a smart city vision. We don’t really have one today. It’s not a product or a service, it’s a framework. And within that framework we will have a series of initiatives that focus on things that are important to us. Transportation is really important to us here in Palo Alto. Energy and resources are really important: we’re going to start to put sensors on important flows of water so we can see the amount of consumption at certain times but also be really smart about leak detection, potentially using little sensors connected to pipes throughout the city. We’re also really focused on the environment. We have a chief sustainability officer who is putting together a multi-decade strategy around what PA needs to do to be part of the solution around climate change.

That’s also going to be a lot about sensors, about collecting data, about informing people and creating positive behaviours. Public safety is another key area. Being able to respond intelligently to crimes, terrorism or natural disasters. A series of sensors again sending information back to some sort of decision system that can help both people and machines make decisions around certain types of behaviours.

How do you expect this whole IoT ecosystem to develop over the next decade?

Bill Gates has a really good saying on this: “We always overestimate the change that will occur in the next two years and underestimate the change that will occur in the next ten.”  It’s something that’s informed me in my thinking. I think things are going to move faster and in more surprising ways in the next ten years for sure: to the extent that it’s very hard to anticipate where things are headed.

We’re disrupting the taxi business overnight, the hotel business, the food business. Things are happening at lightning speed. I don’t know if we have a good sense of where it’s all headed. Massive disruption across all domains, across work, play, healthcare, every sort of part of our lives.

It’s clear that – I can say this – ten years from now won’t be the same as today. I think we’ve yet to see the full potential of smart phones – I think they are probably the most central part of this ongoing transformation.

I think we’re going to connect many more things that we’re saying right now. I don’t know what the number will be: I hear five billion, twenty billion in the next five years. It’s going to be more than that. It’s going to become really easy to connect. We’ll stick a little communication device on anything. Whether it’s your key, your wallet, your shoes: everything’s going to be connected.

Palo Alto and the IoT Data Analytics & Visualization event look like a great matchup. What are you looking forward to about taking part?

It’s clearly a developing area and so this is the time when you want to be acquiring knowledge, networking with some of the big thinkers and innovators in the space. I’m pleased to be part of it from that perspective. Also from the perspective of my own personal learning and the ability to network with great people and add to the body of knowledge that’s developing. I’m going to be kicking it off as the CIO for the city.

Wind River launches comprehensive cloud suite

Cloud computing conceptEmbedded software vendor Wind River has launched what it describes as a ‘comprehensive cloud suite’ for multi-architecture operating systems.

The new Wind River range includes the Helix Cloud, Rocket and Pulsar Linux offerings which are designed to communicate across multiple devices, gateways and microcontroller units (MCUs).

The Helix Cloud is a family of software-as-a-service (SaaS) products including development tools, virtual labs and deployed devices. Their joint mission is to simplify and automate the building and managing of IoT technologies at every stage of the life cycle of a system, from design to decommissioning. The Helix Lab Cloud is a virtual hardware lab for simulating and testing IoT devices and complex systems. Meanwhile, the Device Cloud is designed for managing IoT devices and their data.

Wind River claims it can simplify edge-to-cloud development with a single operating system controlling all dialogue between the device and the cloud. Wind River’s Rocket is described as a tiny-footprint commercial-grade quality real-time operating system that’s directly connected to its Helix Cloud. This, it claims, creates the support for multiple architectures and applications running on the type of 32-bit MCUs used in small-footprint sensor hubs, wearables and edge devices.

Pulsar Linux is a small-footprint commercial-grade binary Linux OS based on the Wind River Linux distribution that connects directly to the Helix Cloud to run on applications scaling from 32-bit MCUs to 64-bit CPUs.

The platform independent Rocket and Pulsar Linux support Intel and ARM architectures and a range of mainstream commercial boards, so that apps can run on any device and the developer can create an open collaborative ecosystem.

Wind River partners include Advantech, Freescale, HCL Technologies, Texas Instruments and Xilinx. It has also launched a new developer programme for ISVs, OEMs, systems integrators, ODMs and cloud operators.

IBM to buy The Weather Company and make it elementary to Watson

IBM The Weather Company PhotoIBM has entered an agreement to buy The Weather Company’s B2B, mobile and cloud-based web properties, in a bid to extend its Internet of Things range.

The acquired assets include WSI, weather.com, Weather Underground and The Weather Company brand. The Weather Channel will not be part of the acquisition but it will license weather forecast data and analytics from IBM under a long-term contract.

IBM says the combination of technology and expertise from the two companies will be foundation for the new Watson IoT Unit and Watson IoT Cloud platform as part of its $3 billion investment strategy in this sector.

The Weather Company’s cloud data system runs the fourth most-used mobile app daily in the United States and handles 26 billion inquiries a day.

On closing the deal, IBM will acquire The Weather Company’s product and technology assets that include meteorological data science experts, precision forecasting and a cloud platform that ingests, processes, analyses and distributes petabyte sized data sets instantly. The Weather Company’s models analyse data from three billion weather forecast reference points, more than 40 million smartphones and 50,000 airplane flights per day, allowing it to offer a broad range of data-driven products and services to 5000 clients in the media, aviation, energy, insurance and government industries.

The Weather Company’s mobile and web properties serves 82 million unique monthly visitors. IBM said it plans to develop The Weather Company’s digital advertising platform and skills, commercialising weather information through data-driven advertising with additional ad-sponsored consumer and business solutions.

“The next wave of improved forecasting will come from the intersection of atmospheric science, computer science and analytics,” said Weather Company CEO David Kenny. “Upon closing of this deal, The Weather Company will continue to be able to help improve the precision of weather forecasts and deepen IBM’s Watson IoT capabilities.”

Microsoft and Azul Systems say Zulu Embedded will encourage IoT in Windows

internet of things farmingAzul Systems and Microsoft are to give Java developers open source development tools, device I/O libraries and a Java runtime targeting Internet of Things (IoT) applications on Windows 10.

The two vendors have created Zulu Embedded for Windows 10 IoT, which is a Java Development Kit (JDK), Java Virtual Machine (JVM) and a set of device I/O libraries. The libraries are based on OpenJDK, which has been certified by Azul for use with Windows 10 IoT Core and is compliant with the Java 8 SE specification.

Microsoft Windows 10 IoT Core, is a modified version of Windows 10 that has been tailored to suit cheap, small-footprint embedded devices such as those based on Raspberry Pi 2 and Minnowboard Max.

The aim of the partnership is to ensure Zulu Embedded meets Java development and runtime requirements for Microsoft’s IoT initiatives. The success of the joint effort will be gauged by the number of Java compatibility updates, security patches and the levels of support for additional IoT device connectivity, control and communication, according to a joint statement.

There are Java developers around the world using Windows 10 IoT core, according to Steve Teixeira, Director of Program Management for the Windows Internet of Things team at Microsoft. These new initiative means they will be assured of a high-quality foundation for their Java projects if they use the latest advances in OpenJDK.

“Developers have many development and deployment choices for their IoT applications,” said Teixeira. By giving them more support, they are more likely to stay in the Microsoft cloud camp, he said. “Microsoft and Azul have made it easy for those who prefer Java to build premier IoT devices running Windows.”

Azul Systems is committed to updating and evolving Zulu Embedded to meet the specific requirements of Microsoft’s IoT platforms, said Scott Sellers, CEO of Azul Systems.

Zulu Embedded for Windows 10 IoT is free to download and use and may be distributed without restriction.

New Teradata apps enable IoT analytics

Internet of things cloudTeradata customers can now listen to the Internet of Things data thanks to two new software innovations designed to create insights into developments.

Teradata’s new Listener and Aster Analytics applications can intelligently listen in real-time and then use analytics to see the distinctive patterns in massive streams of IoT data, it says.

Teradata Listener is an intelligent system that can follow multiple streams of sensor and IoT data wherever it exists globally and feed it to a choice of different analytical systems. Data sent to the Teradata Integrated Big Data Platform 1800 provides access to large volumes of data with its native support of JSON (Java Script Object Notation) data. Alternatively, data fed to a Hadoop based system can be analysed at scale with Teradata Aster Analytics on Hadoop.

Teradata Listener helps data scientists, business analysts and developers to analyse new data streams for faster answers to business questions. Users can analyse data from numerous sources including sensors, telematics, mobile events, click streams, social media feeds and IT server logs, without seeking technical help from the IT department.

Teradata Aster Analytics on Hadoop has 100 pre-configured analytics techniques and seven vertical industry applications to run directly on Hadoop.

In a hospital, data from magnetic resonance imaging (MRI), radiography, and ultrasound imaging equipment might be streamed as text logs. This information describing patient behaviour and sensor data could be streamed into an Hadoop data lake. The new systems allow the users to runs text analytics on the data in order to find out how effectively personnel are working and how efficiently expensive resources, such as MRI scanners, are being used.

“Customers can combine IoT data with business operations and human behavioural data to maximise analytic value,” said Hermann Wimmer, Teradata’s co-president, “Teradata Listener and Aster Analytics on Hadoop are breakthrough IoT technologies that push the analytic edge, making the ‘Analytics of Everything’ possible.”

The collection and analysis of sensor and IOT data has been integral to driving the efficiency of the rail business, according to railways expert Gerhard Kress, director of Analytical Services at Siemens’ Mobility Division.

Kii and KDDI say their joint platform will make IoT safe on cloud

Secure cloudJapanese telco KDDI is working with Internet of Things (IoT) cloud platform provider Kii to create a risk averse system in which enterprises can develop mobile apps.

The KDDI cloud platform service (KCPS) is described as a mobile back end as a service (mBaaS) offering that uses Kii’s software to create mobile and IoT apps on a private network. The two companies have worked together on ways to apply cloud disciplines for efficient sharing of resources, contained within the confines on an Intranet environment. The object of the collaboration is to allow companies to develop machine to machine systems, without exposing them to the public cloud while they are in development.

According to KDDI, the KCPS uses the telco’s Wide Area Virtual Switch to integrate a number of different virtual network layers with Kii’s software. Together they create a new level of fast connections across the Intranet. KCPS also provides a service environment for intranet-conscious customers who need high standards of security and enterprise functions without resorting to the public Internet, according to the vendor.

KDDI claims this is the first instance in which both Intranet and Internet services can work seamlessly with any mobile application developed on the KCPS platform.

KDDI’s application development support will allow developers to build better quality, lower priced applications in a short period of time, it claims. The platform is designed to help developers manage application development, devices and data, while providing essential features like push notifications and geo-location information. KCPS should be compatible with mobile apps on Android and iOS, according to KDDI.

“As the IoT gains mass acceptance, we see tremendous value helping mobile app developers get more IoT devices into the hands of consumers,” said Kii CEO Masanari Arai, “our collaboration will use the cloud to build the backend support of these apps in Japan.”

Forecasting Cloud’s Place in the Internet of Things

image (1)

 

Date: Thursday, October 22, 2015

Time: 04:00 PM British Summer Time

Duration: 1 hour

Data is no longer contained within the walls of a data center. Devices or “things” need to share data with the same quality of service whether they are in remote corners of the world or in major metropolitan areas. Flexibility, speed, and security area all required elements for real-time data analysis in IoT projects and any company wanting to transform with the Internet of Things; as a result, the Internet of Things is becoming the primary application for cloud computing.

In a recent study from IBM and Business Cloud News, we learned that nearly two-thirds of respondents agree companies that are slow to integrate cloud into their IoT solutions will fall behind the competition. Learn more about current trends and future growth opportunities as data, engagement, and cloud all converge around the Internet of Things.

  • See how the Internet of Things is becoming the primary application for cloud computing
  • Review your readiness to integrate cloud and IoT solutions

REGISTER FOR THIS WEBINAR

MapR claims JSON IoT development breakthrough

Cloud databaseEnterprise software vendor MapR has unveiled plans to slash the workload of IoT developers and administrators by cutting the complexity of managing its NoSQL databases.

The key to this simplification, it says, is in more creative use of the JavaScript Object Notification (JSON) format, which it claims has the potential to make significant improvements in both database scalability and the analysis of the information they contain.

“We’re seeing big changes in the way applications are developed and how data is consumed,” said MapR’s chief marketing officer Jack Norris, “the underlying data format is the key to making information sharing easier.”

Bringing out the advantages of JSON makes administration easier, according to Norris, because users can make changes easily in a database built on documents. This in turn helps developers when they are planning applications, because it is easier to create a user friendly system. Tweaking JSON will benefit system builders in their own work too, Norris argued, since a document database can now be given enterprise grade scalability, reliability and integrated analytics.

The organisational improvements include the ability to personalise and deliver better online shopping experiences, reduce risk and prevent fraud in real-time, improve manufacturing efficiencies and cut costs. Savings will come from preventing cluster sprawl, eliminating data silos and lowering the cost of ownership of data management, claims MapR. Meanwhile it has promised a productivity dividend from continuous analytics of real-time data.

The MapR-DB supports the Open JSON Application Interface (OJAITM), which is designed to be a general purpose JSON access layer across databases, file systems and message streams, enabling a flexible and unified interface to work with big data, claims MapR.

The addition of a document database capacity extends the NoSQL MapR-DB to cover more types of unstructured business data, said one analyst. This could make it faster and easier to build big data applications, without the burden of shuffling data around first.

“MapR continues to build on the innovative data platform at the core of its Hadoop distribution,” said Nik Rouda, senior analyst at the Enterprise Strategy Group.

How IoT Security could change infrastructure forever

CybersecurityOn September 22nd and 23rd, the first-ever dedicated IoT Security conference and exhibition will take place in Boston.

While at first glance this may appear to concern a specific and rather specialized area, the relationship of the Internet of Things to the broad issue of human security may well prove much more far-reaching and fundamental.

After all, the development of the Internet itself was driven by a Cold War desire to create resilient computer networks that could withstand a nuclear attack. This threat inspired a whole new architecture for sharing and protecting information – one that was intentionally decentralized.

History suggests that precaution can be a key driver of technological innovation. In changing things to protect them, we often open up unforeseen new opportunities.

Which is why, if we return to 2015, there is something fascinating in seeing the same decentralized architectures applied to real-world infrastructures in the name of collective safety.

“When you apply this kind of Internet-type architecture to core infrastructure — whether it’s water or energy or transportation –  these systems start looking a lot more like the Internet,” says John Miri, Chief Administrative Officer at the Lower Colorado River Authority (LRCA) and a speaker at this month’s Boston event. “You start to see water systems, flood data systems and, hopefully, electric grids that are less centralized, more resilient and more difficult to disrupt.”

The LCRA is an 80-year-old institution with roots in the Great Depression, entrusted with providing reliable water, flood protection and electricity to Central Texas and beyond. The areas LCRA serves covers a number of the fastest growing cities in the United States, meaning LCRA faces some pretty substantial demands on its infrastructure.

“Providing the water and power to support growing communities and a growing business and industrial base is no small task,” Miri says. Indeed, LCRA has  broken ground on a quarter of a billion dollar new reservoir, the region’s first new water supply in decades.

Many of these additional demands make  safety and security more important than ever.

“LCRA is now the second largest electric transmission utility in Texas. Our high tension transmission lines go across a large portion of the state. Protecting the electric grid is a pretty hot topic,” Miri says.

These hypothetical threats encompass what Miri calls “bad actors,”  but also less hypothetical threats to the infrastructure.

“When you have a flood, we may have to intentionally shut down electric substations. Everyone knows electricity and water don’t mix – but even having the situational awareness to know that water is approaching a substation is very important to us in keeping the lights on. Using these kinds of smart networks to get a better picture of the threats and dangers to the power grid helps us protect it rather than just saying ‘build more,’” Miri says.

Similarly, a vast number of sensors throughout its Hydromet network enable LCRA to better monitor water levels – and to effectively manage floods.

“By adopting a new, more open, shared technology approach, we could expand the infrastructure we have for flood data collection at a 90% lower cost than if we had done it a traditional way. The technology  actually opens up our infrastructure to a very wide region that never considered it before. We can offer a level of flood monitoring across a wider region and  extend it rural and agricultural communities and other areas that might not have the resources to gain access to this technology.”

Looking ahead, Miri says, there are new opportunities to apply this decentralized, Internet-style architecture to other projects.

“I think when you look forward 10, 15 or 20 years, the whole infrastructure may work differently. It opens up new possibilities and business models that we didn’t have before. For instance, Texas is on the coast. As with any coastal area, we spend time thinking about desalination. Some of the work we’ve been doing on the Internet of Things  is making people think, maybe we don’t need a couple of giant desalination plants – which has been the approach in Australia and Israel – but a number of smaller plants that are networked together, and share the water more efficiently. In the longer term, IoT may actually change the infrastructure itself, which would be very exciting.”

It could be interesting to one day look back at this month’s inaugural IoT Security event and see how many of the topics discussed went on to fundamentally evolve and affect their wider respective domains.

Salesforce IoT Cloud promises to create meaning from M2M talk

Internet of Things flat iconic illustrationSalesforce has launched its Internet of Things (IoT) offering with IoT Cloud, which promises to convert machine to machine conversations, digital content and customer information into useful intelligence that sales staff can act on.

The IoT Cloud is powered by event processing engine Salesforce Thunder, which will allow customers to personalise the way they sell, service and market. Development partners include processor maker ARM, Etherios, Informatica, PTC ThingWorx and Xively LogMeln.

The service was unveiled by Salesforce CEO Marc Benioff at the company’s Dreamforce conference in San Francisco. “Salesforce is turning the internet of things into the internet of customers,” said Benioff. The IoT Cloud will allow businesses to create instant one-to-one proactive actions for sales, service, marketing or any other business process, Benioff said.

According to Salesforce, its ‘massively scalable’ cloud computing architecture can ‘listen’ to the connected world and make sense of the billions of events each day from all sources. The connections with wearables, windmills, telephones and turbines – and all other devices – can be contextualized by Salesforce’s own real-time rules, it claims. The IoT Cloud aims to give business users intuitive, point-and-click tools to define rules and logic for events that can trigger actions among the users of Salesforce’s customer relationship systems.

The rationale is to glean information reported by devices – such as the numbers of hard braking movements by drivers of a car fleet – in order to monitor and manage customer cases. Machine intelligence, such as vehicle braking data, could help Salesforce users negotiate from a position of superior knowledge with their customers.

According to Salesforce’s figures there will be 75 billion devices connected to the Internet by 2020, with the volume of data growing exponentially each year. The McKinsey Global Institute estimates that IoT applications have the potential to make $11.1 trillion worth of economic impact per year by 2025.

The challenge is to make sense of all that data, said IoT market watcher Gary Barnett, chief analyst at Ovum. “IoT deployments only bring value when organisations can act on the information their IoT networks generate,” said Barnett, chief analyst, Ovum. “The ability to make sense of that data will be a key factor in turning it into action.”