- Modeling: Another key element is the modeling that is used to define how the data will be processed to automatically create accurate predictive models, Rajagopalan says. The algorithms can be as simple as rules that can be applied to understand a particular situation or understand data in the context of a particular scenario. There are also supervised algorithms and models that use machine learning techniques to build hypotheses around trends in the data and constantly refine themselves based on the data they are presented with.
- Deployment: IT leaders then have the outputs of the model, such as a visualization, report or chart. The results of the predictive analysis are then given to decision-makers.
- Model monitoring: The models are continuously monitored to ensure they are providing the results that are expected.
Before, Rajagopalan says, agencies had specialized units to apply SAS, but those models were expensive to create. The democratization and consumerization of data and of analytics tools has made it easier to create simple and succinct summaries of data that visualize outputs.
READ MORE: Find out how agencies can unleash the power of data analytics.
What Is Open Data?
Joshua New, formerly a policy analyst at the Center for Data Innovation and now a technology policy executive at IBM, tells FedTech that open data is best thought of as “machine-readable information that is freely available online in a nonproprietary format and has an open license, so anyone can use it for commercial or other use without attribution.”
On May 9, 2013, former President Barack Obama signed an executive order that made open and machine-readable data the new default for government information.
“Making information about government operations more readily available and useful is also core to the promise of a more efficient and transparent government,” the Obama administration noted.
On Jan. 14, 2019, the OPEN Government Data Act, as part of the Foundations for Evidence-Based Policymaking Act, became law. The OPEN Government Data Act makes data.gov a requirement in statute, rather than a policy. It requires agencies to publish their information online as open data, using standardized, machine-readable data formats, with their metadata included in the data.gov catalog. May 2019 marks the 10th anniversary of data.gov, the federal government’s open data site.
The General Services Administration launched the site with a modest 47 data sets, but the site has grown to over 200,000 data sets from hundreds of data sources including federal agencies, states, counties and cities. “Data.gov provides easy access to government datasets covering a wide range of topics — everything from weather, demographics, health, education, housing, and agriculture,” according data.gov.
MORE FROM FEDTECH: See how agencies can get the most value out of their data.
What are Examples of Open Data?
There are numerous federal open data programs. FarmPlenty helps farmers better analyze Agriculture Department open data on crops grown within a 5-mile radius of their farms. The application that supports the program was built as part of the USDA-Microsoft Innovation Challenge and is supported by the USDA’s National Agricultural Statistics Service CropScape and Quickstats application programming interfaces.
Where are the Jobs? is an app that uses data from the Census Bureau and the Bureau of Labor Statistics to allow users to interactively explore the salary and job statistics for various occupations at national, state and regional levels. Home Energy Saver is an interactive consumer application used to estimate residential energy use and plan home energy efficiency upgrades that uses open data from the Energy Information Administration.
Predictive Analytics Examples in Government
Federal agencies are using predictive analytics for a wide range of use cases, including cybersecurity. Specifically, agencies are using these tools to predict insider threats, Splunk’s Jardim says. The models look at users’ backgrounds, where they have worked, how often they have logged in to networks at certain times and whether that behavior actually is anomalous. The goal of such tools is to make a good prediction of whether the security events should be tracked by human analysts, Jardim says.
“You only want to surface the events that are very clear insider threats,” he says. “The analyst is focused on high-probability events, not low-probability events.”
Predictive analytics can also be used for agencies’ data center maintenance by applying algorithms to look at compute capacity, how many users are accessing services and to assess throughput for mission-critical applications, Jardim says. Such tools can predict when a particular server will become overloaded and can help agencies preempt those events to ensure users have access to vital applications.
The Defense Department can also use predictive analytics to ensure that soldiers have enough of the right munitions and supplies in particular theaters of war and enough support logistics. “Logistics and operational maintenance take on a life-or-death consequence if I cannot ship enough munitions or vehicles into a specific theater,” Jardim says.
Qlik’s Churchill says that a customer within the Army is using predictive analytics tools to build models that support force enablement and predict the capabilities that will be needed in the future and which capabilities will be diminished, as well as the capabilities that will be required if certain scenarios arise.
The Pentagon is also working on predictive analytics tools for financial management via the Advanta workflow tool, which has brought together roughly 200 of the DOD’s enterprise business systems, Churchill says.
“How can they use predictive models to understand the propensity to have to de-obligate funds from a particular program in the future?” Churchill says. “As I am evaluating the formulation and execution of budgets, technologies like this have the ability to help those decision-makers identify the low-hanging fruit. How do I put those insights in front of people that they wouldn’t have gotten before?”
Predictive maintenance is also a key use case, especially for vehicles and other heavy equipment. Models can ingest data such as the weather and operating conditions of vehicles, not just how many hours they have been running, to determine when they will break down, Churchill says.