13 January 2020

Using Machine Learning to “Nowcast” Precipitation in High Resolution




The weather can affect a person’s daily routine in both mundane and serious ways, and the precision of forecasting can strongly influence how they deal with it. Weather predictions can inform people about whether they should take a different route to work, if they should reschedule the picnic planned for the weekend, or even if they need to evacuate their homes due to an approaching storm. But making accurate weather predictions can be particularly challenging for localized storms or events that evolve on hourly timescales, such as thunderstorms.

In “Machine Learning for Precipitation Nowcasting from Radar Images,” we are presenting new research into the development of machine learning models for precipitation forecasting that addresses this challenge by making highly localized “physics-free” predictions that apply to the immediate future. A significant advantage of machine learning is that inference is computationally cheap given an already-trained model, allowing forecasts that are nearly instantaneous and in the native high resolution of the input data. This precipitation nowcasting, which focuses on 0-6 hour forecasts, can generate forecasts that have a 1km resolution with a total latency of just 5-10 minutes, including data collection delays, outperforming traditional models, even at these early stages of development.

Moving Beyond Traditional Weather Forecasting
Weather agencies around the world have extensive monitoring facilities. For example, Doppler radar measures precipitation in real-time, weather satellites provide multispectral imaging, ground stations measure wind and precipitation directly, etc. The figure below, which compares false-color composite radar imaging of precipitation over the continental US to cloud cover imaged by geosynchronous satellites, illustrates the need for multi-source weather information. The existence of rain is related to, but not perfectly correlated with, the existence of clouds, so inferring precipitation from satellite images alone is challenging.
Top: Image showing the location of clouds as measured by geosynchronous satellites. Bottom: Radar image showing the location of rain as measured by Doppler radar stations. (Credit: NOAA, NWS, NSSL)
Unfortunately, not all of these measurements are equally present across the globe. For example, radar data comes largely from ground stations and is generally not available over the oceans. Further, coverage varies geographically, and some locations may have poor radar coverage even when they have good satellite coverage.

Even so, there is so much observational data in so many different varieties that forecasting systems struggle to incorporate it all. In the US, remote sensing data collected by the National Oceanic and Atmospheric Administration (NOAA) is now reaching 100 terabytes per day. NOAA uses this data to feed the massive weather forecasting engines that run on supercomputers to provide 1- to 10-day global forecasts. These engines have been developed over the course of the last half century, and are based on numerical methods that directly simulate physical processes, including atmospheric dynamics and numerous effects like thermal radiation, vegetation, lake and ocean effects, and more.

However, the availability of computational resources limits the power of numerical weather prediction in several ways. For example, computational demands limit the spatial resolution to about 5 kilometers, which is not sufficient for resolving weather patterns within urban areas and agricultural land. Numerical methods also take multiple hours to run. If it takes 6 hours to compute a forecast, that allows only 3-4 runs per day and resulting in forecasts based on 6+ hour old data, which limits our knowledge of what is happening right now. By contrast, nowcasting is especially useful for immediate decisions from traffic routing and logistics to evacuation planning.

Radar-to-Radar Forecasting
As a typical example of the type of predictions our system can generate, consider the radar-to-radar forecasting problem: given a sequence of radar images for the past hour, predict what the radar image will be N hours from now, where N typically ranges from 0-6 hours. Since radar data is organized into images, we can pose this prediction as a computer vision problem, inferring the meteorological evolution from the sequence of input images. At these short timescales, the evolution is dominated by two physical processes: advection for the cloud motion, and convection for cloud formation, both of which are significantly affected by local terrain and geography.
Top (left to right): The first three panels show radar images from 60 minutes, 30 minutes, and 0 minutes before now, the point at which a prediction is desired. The right-most panel shows the radar image 60 minutes after now, i.e., the ground truth for a nowcasting prediction. Bottom Left: For comparison, a vector field induced from applying an optical flow (OF) algorithm for modeling advection to the data from the first three panels above. Optical flow is a computer vision method that was developed in the 1940s, and is frequently used to predict short term weather evolution. Bottom Right: An example prediction made by OF. Notice that it tracks the motion of the precipitation in the bottom left corner well, but fails to account for the decaying strength of the storm.
We use a data-driven physics-free approach, meaning that the neural network will learn to approximate the atmospheric physics from the training examples alone, not by incorporating a priori knowledge of how the atmosphere actually works. We treat weather prediction as an image-to-image translation problem, and leverage the current state-of-the-art in image analysis: convolutional neural networks (CNNs).

CNNs are usually composed of a linear sequence of layers, where each layer is a set of operations that transform some input image into a new output image. Often, a layer will change the number of channels and the overall resolution of the image it’s given, in addition to convolving the image with a set of convolutional filters. These filters are themselves small images (for us, they are typically only 3x3, or 5x5). Filters drive much of the power of CNNs, and result in operations like detecting edges, identifying meaningful patterns, etc.

A particularly effective type of CNN is the U-Net. U-Nets have a sequence of layers that are arranged in an encoding phase, in which layers iteratively decrease the resolution of the images passing through them, and then a decoding phase in which the low-dimensional representations of the image created by the encoding phase are expanded back to higher resolutions. The following figure shows all of the layers in our particular U-Net.
(A) The overall structure of our U-NET. Blue boxes correspond to basic CNN layers. Pink boxes correspond to down-sample layers. Green boxes correspond to up-sample layers. Solid lines indicate input connections between layers. Dashed lines indicate long skip connections transversing the encoding and decoding phases of the U-NET. Dotted lines indicate short skip connections for individual layers. (B) The operations within our basic layer. (C) The operations within our down-sample layers. (D) The operations within our up-sample layers.
The input to the U-Net is an image that contains one channel for each multispectral satellite image in the sequence of observations over the last hour. For example, if there were 10 satellite images collected in the last hour, and each of those multispectral images was taken at 10 different wavelengths, then the image input for our model would be an image with 100 channels. For radar-to-radar forecasting, the input is a sequence of 30 radar observations over the past hour, spaced 2 minutes apart, and the output contains the prediction for N hours from now. For our initial work in the US, we trained a network from historical observations over the continental US from the period between 2017 and 2019. The data is split into periods of four weeks, where the first three weeks of each period are used for training and the fourth week is used for evaluation.

Results
We compare our results with three widely used models. First, the High Resolution Rapid Refresh (HRRR) numerical forecast from NOAA. HRRR actually contains predictions for many different weather quantities. We compared our results to their 1-hour total accumulated surface precipitation prediction, as that was their highest quality 1-hour precipitation prediction. Second, an optical flow (OF) algorithm, which attempts to track moving objects through a sequence of images. This latter approach is often applied to weather prediction even though it makes the assumption that overall rain quantities over large areas are constant over the prediction time — an assumption that is clearly violated. Third, the so-called persistence model, is the trivial model in which each location is assumed to be raining in the future at the same rate it is raining now, i.e. the precipitation pattern does not change. That may seem like an overly simplistic model to compare to, but it is common practice given the difficulty of weather prediction.
A visualization of predictions made over the course of roughly one day. Left: The 1-hour HRRR prediction made at the top of each hour, the limit to how often HRRR provides predictions. Center: The ground truth, i.e., what we are trying to predict. Right: The predictions made by our model. Our predictions are every 2 minutes (displayed here every 15 minutes) at roughly 10 times the spatial resolution made by HRRR. Notice that we capture the general motion and general shape of the storm.
We use precision and recall (PR) graphs to compare the models. Since we have direct access to our own classifier, we provide a full PR curve (seen as the blue line in the figure below). However, since we don’t have direct access to the HRRR model, and since neither the persistence model nor OF have the ability to trade-off precision and recall, those models are represented only by individual points. As can be seen, the quality of our neural network forecast outperforms all three of these models (since the blue line is above all of the other model’s results). It is important to note, however, that the HRRR model begins to outperform our current results when the prediction horizon reaches roughly 5 to 6 hours.
Precision and recall (PR) curves comparing our results (solid blue line) with: optical flow (OF), the persistence model, and the HRRR 1-hour prediction. As we do not have direct access to their classifiers, we cannot provide a full PR curve for their results. Left: Predictions for light rain. Right: Predictions for moderate rain.
One of the advantages of the ML method is that predictions are effectively instantaneous, meaning that our forecasts are based on fresh data, while HRRR is hindered by computational latency of 1-3 hours. This leads to better forecasts for computer vision methods for very short term forecasting. In contrast, the numerical model used in HRRR can make better long term predictions, in part because it uses a full 3D physical model — cloud formation is harder to observe from 2D images, and so it is harder for ML methods to learn convective processes. It's possible that combining these two systems, our ML model for rapid forecasts and HRRR for long-term forecasts, could produce better results overall, an idea at the focus of our future work. We're also looking at applying ML directly to 3D observations. Regardless, immediate forecasting is a key tool for real-time planning, facilitating decisions and improving lives.

Acknowledgements
Thanks to Carla Bromberg, Shreya Agrawal, Cenk Gazen, John Burge, Luke Barrington, Aaron Bell, Anand Babu, Stephan Hoyer, Lak Lakshmanan, Brian Williams, Casper Sønderby, Nal Kalchbrenner, Avital Oliver, Tim Salimans, Mostafa Dehghani, Jonathan Heek, Lasse Espeholt, Sella Nevo, Avinatan Hassidim.

Quibi execs Jeffrey Katzenberg and Meg Whitman explain their big vision


Last week at the Consumer Electronics Show in Las Vegas, Quibi executives — including CEO Meg Whitman and founder/chairman Jeffrey Katzenberg — took the stage in a keynote laying out their vision for the mobile video service.

Katzenberg is a longtime Hollywood executive who led Walt Disney Studios during its animation renaissance in the late ’80s and early ’90s before co-founding Dreamworks Animation. Whitman worked at both Disney and Dreamworks, but she’s best known as the former CEO of eBay and Hewlett Packard Enterprise.

So it’s fitting that they presented Quibi as a company that exists at the intersection of Hollywood and Silicon Valley — as Whitman put it, creating “the very first entertainment technology platform optimized for mobile viewing.”


Read Full Article

Google Cloud launches new solutions for retailers


It’s no secret that the Google Cloud management team has decided to focus its efforts on a select number of enterprise verticals like healthcare, manufacturing, financial services, energy and life sciences. Retail, too, has long been a growth market for the company, especially as Amazon’s competitors are looking to run their services on clouds that are not AWS. Current customers include the likes of Kohl’s, Lowe’s and France’s Carrefour. It’s maybe no surprise then that Google today used NRF 2020, one of the largest retail events, to launch a number of updates to its services for retailers.

Some of the announcements today focus on specific vertical editions of existing services, including Google Cloud API Management for Retail, powered by Apigee, or Google Cloud Anthos for Retail, which specifically targets retailers that want to modernize their store operations and infrastructure. There is also Google Cloud Search for Retail, powered by Google Search, which promises to bring better product search results to a retailer’s applications.

In addition, Google is also expanding programs like its Retail Acceleration Program to more customers and making its white-glove Customer Reliability Engineering service, which helps retailers better plan for and manage their peak shopping days, available to more customers.

What’s maybe more interesting, though, is new services like Google Cloud 1:1 Engagement for Retail, “a blueprint and best-practice guide on how to build these types of data-driven solutions effectively and with less up-front cost.” The idea here is to help retailers make use of Google’s big data platform to build personalization and recommendation models to better understand and engage their customers.

Also new is a buy optimization and demand forecasting service that aims to help retailers better plan their logistics operations.

We’ll likely see Google use a similar playbook for more verticals over time. We know that Google Cloud has ambitions to become the #2 cloud within a few years and to do so, it needs to get large enterprises — and especially those that are still trying to figure out their cloud strategies — to opt for its services.


Read Full Article

Why "biofabrication" is the next industrial revolution | Suzanne Lee

Why "biofabrication" is the next industrial revolution | Suzanne Lee

What if we could "grow" clothes from microbes, furniture from living organisms and buildings with exteriors like tree bark? TED Fellow Suzanne Lee shares exciting developments from the field of biofabrication and shows how it could help us replace major sources of waste, like plastic and cement, with sustainable and eco-friendly alternatives.

Click the above link to download the TED talk.

Samsung acquires TeleWorld Solutions to help build 5G infrastructure


Samsung this morning announced that it has completed the acquisition of TeleWorld Solutions. The Virginia-based telecommunications company provides wireless networking and consulting services. It’s TWS’s 5G solutions that Samsung is clearly the most interested in as part of this deal.

The electronics giant says it plans to leverage TWS’s services to help U.S.-based networks build out the next generation wireless.

“The acquisition of TWS will enable us to meet mobile carriers’ growing needs for improving their 4G and 5G networks, and eventually create new opportunities to enhance our service capabilities to our customers,” Samsung EVP Paul Kyungwhoon Cheun said in a release. “Samsung will continue to drive innovation in communications technology, while providing optimization services for network deployments that accelerate U.S. 5G network expansion.”

The deal will make TWS a wholly owned subsidiary of Samsung, allowing the brand to continue to offer its consulting services to existing clients. That last bit is important, so as to not leave companies in a lurch over the course of the next year, as 5G becomes an increasing focus beyond just smartphone connectivity.


Read Full Article

Former Google Pay execs raise $13.2M to build neo-banking platform for millennials in India


Two co-founders of Google Pay in India are building a neo-banking platform in the country — and they have already secured backing from three top VC funds.

Sujith Narayanan, a veteran payments executive who co-founded Google Pay in India (formerly known as Google Tez), said on Monday that his startup, epiFi, has raised $13.2 million in its Seed financial round led by Sequoia India and Ribbit Capital. The round valued epiFi at about $50 million.

David Velez, the founder of Brazil-based neo-banking giant Nubank, Kunal Shah, who is building his second payments startup CRED in India, and VC fund Hillhouse Capital also participated in the round.

The eight-month-old startup is working on a neo-banking platform that will focus on serving millennials in India, said Narayanan, in an interview with TechCrunch.

“When we were building Google Tez, we realized that a consumer’s financial journey extends beyond digital payments. They want insurance, lending, investment opportunities and multiple products,” he explained.

The idea, in part, is to also help users better understand how they are spending money, and guide them to make better investments and increase their savings, he said.

At this moment, it is unclear what the convergence of all of these features would look like. But Narayanan said epiFi will release an app in a few months.

Working with Narayanan on epiFi is Sumit Gwalani, who serves as the startup’s co-founder and chief product and technology officer. Gwalani previously worked as a director of product management at Google India and helped conceptualize Google Tez. In a joint interview, Gwalani said the startup currently has about two-dozen employees, some of whom have joined from Netflix, Flipkart, and PayPal.

Shailesh Lakhani, Managing Director of Sequoia Capital India, said some of the fundamental consumer banking products such as savings accounts haven’t seen true innovation in many years. “Their vision to reimagine consumer banking, by providing a modern banking product with epiFi, has the potential to bring a step function change in experience for digitally savvy consumers,” he said.

Cash dominates transactions in India today. But New Delhi’s move to invalidate most paper bills in circulation in late 2016 pushed tens of millions of Indians to explore payments app for the first time.

In recent years, scores of startups and Silicon Valley firms have stepped to help Indians pay digitally and secure a range of financial services. And all signs suggest that a significant number of people are now comfortable with mobile payments: More than 100 million users together made over 1 billion digital payments transaction in October last year — a milestone the nation has sustained in the months since.

A handful of startups are also attempting to address some of the challenges that small and medium sized businesses face. Bangalore-based Open, NiYo, and RazorPay provide a range of features such as corporate credit cardsa single dashboard to manage transactions and the ability to automate recurring payouts that traditional banks don’t currently offer. These platforms are also known as neo-bank or challenger banks or alternative banks. Interestingly, most neo-banking platforms in South Asia today serve startups and businesses — not individuals.


Read Full Article