A comprehensive model for management and validation of federal big data analytical systems
Big Data Analytics volume 2, Article number: 2 (2017)
In this era of data science, many software vendors are rushing towards providing better solutions for data management, analytics, validation and security. The government, being one of the most important customers, is riding the wave of data and business intelligence. However, federal agencies have certain requirements and bureaucracies for data-related processes, certain rules and specific regulations that would entail special models for building and managing data analytical systems. In this paper, and based on work done at the US government, a model for data management and validation is introduced: Federal Model for Data Management and Validation (FedDMV). FedDMV is 4-step model that has a set of best practices, databases, software tools and analytics. Automated procedures are used to develop the system and maintain it, and association rules are used for improving its quality.
After working with multiple engineers and analysts at the federal agency, there is a general consent that FedDMV is easy to follow (please refer to the experimental survey). However, to quantify that satisfaction, three experimental studies were performed. One is a comparison to other state-of-the-art development models at the government, the second one is a survey that was collected at the government to quantify the level of satisfaction regarding FedDMV and its tool; and finally, a data validation study was performed through detailed testing of the federal system (using an Association Rules algorithm).
To develop a safe and sound federal data analytical system, a tested and rigorous model is required. There is a lack of government-specific models in industry and research. FedDMV aims to provide solutions and guided steps to facilitate the development of data analytics systems given the governmental constraints. FedDMV deals with unstructured data that streams from multiple sources, automates steps that are usually manual, validates the data and maximizes its security. The results of the experimental work are recorded and reported in this manuscript.
The promise of intelligent and accurate predictions in software systems that was previously pursued by many [1, 2] is now being transformed into a new testament of big data analytics. Undoubtedly, data analytics is a buzz word, if not the “buzziest” word of the day [3–6]; it is a science that lies in the nexus of Statistics and Artificial Intelligence (AI). The goal of data analytics is to return an intelligent and more focused version of large and impersonal datasets, provide quick insights into data, and help with visualization and decision making. Literature has many examples of successful applications of data analytics, not only to specific business-driven functions, but also many industrial and research domains such as Healthcare , Education , Banking & Finance , and many others. The Government is no different. Actually, most federal agencies understand that data can help them unlock the government of the future towards better operations, better citizen service, and more efficient decision making.
Data engineering at the US government
In 2012, the Big Data Initiative  (refer to the screenshot shown in Fig. 1 below) was created; that was driven by an earlier Open Government initiative signed in 2009 [8–10]. These two initiatives aimed to create data openness and ensure machine-readable data at all federal agencies. Accordingly, multiple agencies started sharing their data through outlets such as www.data.gov. The Open Data Initiative announcement’s was reinforced in 2013 as an executive order.
The two initiatives created a new reality within federal agencies: an instant need to implement a data system that would allow for managing, browsing, mining, and sharing their data. In this manuscript, a data system is defined as:
A group of databases
Tools that manage the databases
A set of validation, security and safety procedures
The outcomes of the data system such as dashboards and visualizations.
Based on this novel direction, federal agencies are racing to implement data analytical systems, this paper aims to guide these agencies through the process – using the proposed federal-specific model. Although the private sector is known to be the main driver of technological advancements, the increasing trend of software engineering in the government is undeniable , and is driving the change as well. Previously, software development at the government had followed the generic waterfall model , where everything is sequential. That incurred massive costs and time delays. Based on an article by the Washington Post , a monetary value of up to $50 million was wasted on failing federal software projects (in one year). Many governmental agencies (such as: the Financial Industry Regulatory Authority and the Department of Health and Human Services, and many others) still latch onto the expensive and difficult to manage waterfall model (the Affordable Care Act website www.healthcare.gov is a recent example ). Overall, federal agencies spend around $77 billion dollars on software and IT every year . Based on that, the United States Government Accountability Office (GAO) published an important report  making the case for the need to switch government software engineering practices to agile . More importantly, the GAO report identified 14 challenges that governmental agencies undergo. Showed in (Table 1) below, the challenges are specific to governmental workflows; most challenges are associated with Collaboration, Teamwork, and old fashioned federal practices. This fast movement towards agile software development and data analytics is the major motivation for our work.
In a thesis/study published at Princeton University, data structuring was identified as one of the major challenges at federal agencies [17, 18]: “When government does collect and publish data in a reusable way, government enables third-party stakeholders like advocates, academics, journalists and others to powerfully adapt its data in any way they see fit using the latest technologies, and to add value in unexpected ways. Third parties can use government data to experiment in parallel, in order to discover what innovations work best in changing technological environments”. To engineer better data systems however, it is important to first understand the nature of data available in the government; the next section discusses existing and previous federal experiences in this realm.
This paper is structured as follows: the next sub-section looks into what has been already done in the government in terms of software engineering and big data analytics. The section after discusses related work and current federal challenges. Afterwards, the main contribution of this paper, the Federal Data Management and Validation Model (FedDMV) is presented. Subsequently, two experimental studies and survey (on the usability of the tool) are introduced; and the last section concludes the paper with results and future work.
Unstructured data at the US government
To address a legislative issue, for decision making, and for media announcements, most federal agencies require data that streams from other federal departments; or in some cases even city, county or state governments. Some of that data is shared with the public.
Besides, the aforementioned www.data.gov, the PACER system (public courts online access system) , the BLS (Bureau of Labor Statistics), and NASS (National Agricultural Statistics Service)  are major examples of federal data system that adopted the open data initiative and made their data public to citizens. States and cities started publishing data to the public as well. For example, the cities of Chicago  and San Francisco  both have open data portals, data.sfgov.org and data.cityofchicago.org respectively. However, different parties share and store the data in different formats, varying standards, and using different technologies. How is federal data used collectively then? How is it shared across agencies and governments?
Due to the size, unstructured and non-repetitive nature of federal data, Hadoop could be very helpful; Swish Data (among others such as Map Reduce and Horton Works) [22–24] published an article identifying multiple use cases for the government using Hadoop. Besides the ability to organize enormous amounts of size of data, Hadoop also handles unstructured data (which is a major characteristic of federal data). The government is indeed focused on using commercial data management tools such as: Microsoft’s SQL servers , SAS , and Tableau  (among others). Nonetheless, and regardless of the tool of choice, following a controlled data management model reduces the risk of building faulty data systems; it also ensures the ability to learn from other experiences and not repeat the mistakes of the past  (among many other advantages). It is apparent that big data is key for the US government (Table 1 and Fig. 2 illustrate use cases that show why big data is an inevitable reality at the government), however, what tools is the government using to manage its data? Besides rational databases, SQL, Oracle tools and the traditional Database Management Systems (DBMS), is the government using the de-facto of big data – Hadoop? Do they have specific tools that can manage and adhere to federal rules? No strong evidence of that is reported.
Related work: data management models
Lifecycles in software engineering have lead major shifts in the tech progress of the engineering world. The transition from Waterfall development to Spiral , from Traditional to Agile  and from product-based to contextual user-based [29, 30] development left an obvious fingerprint on the world of engineering. Do these models apply to developing a data analytics system? Although considered a fairly novel field, this section introduces the most prominent models for data analytical systems and how they are utilized.
Data analytics development for non-federal systems
Before federal got on the analytical wave, many industries (such as: healthcare, finance, athletics, and the media) adopted analytical models within their organizations. Although data mining research has been of interest to many researchers around the world, data analytics didn’t see much light until it was adopted by the industry. Many software vendors (SAS, SPSS, Tableau, MSTR, Qlik, and Pentaho) shifted the focus of their software development to include a form of data analytics, big data, data mining, statistical modeling and data visualization. Based on multiple long and challenging deployments in many fields, trials and errors, and multiple consulting exchanges with many customers from many fields, these vendors coined a data management model for data analytics. SAS (based on Gartner’s research  is one of the pioneer vendors in this field), provided that model (illustrated in Fig. 3). The model includes the following steps:
Identify and formulate the problem
Prepare the data (pivoting and data cleansing)
Data exploration (summary statistics, bar charts and other means of exploration)
Data transformation and selection (select ranges, and subsets)
Statistical model development,
Validation and deployment
Evaluate and monitor results of model with data
Deliver and refine the model.
When software vendors like Tableau, SAS, SPSS and others realized that the government is highly interested in data analytics as well, they started to propose different models to the government to win its projects. Next section discusses some of these proposed models, and later sections of the paper introduce our new model: The Federal Data Management and Validation Model (FedDMV), and contrast it with an existing relevant model.
Federal data management, validation and analytics
As it is already established, federal departments and agencies are very good with generating and collecting data, but not as good in storing and sharing it. Government IT however, is being constantly challenged to make this data available to employees, the media, the public and other agencies. To address such challenges, multiple software vendors produced federal-specific solutions. Actuate  for example introduced BIRT (a data driven model). It is one of the most used models in government. It has been used by more than million users and features an active community, as Actuate claims . The BIRT process matches most federal regulations and certifications. BIRT provides a list of reports and dashboards that agencies can plug into their data; it has dashboards that deal with federal data sharing, data fraud detection, performance management, and citizen self-service. Actuate deploys BIRT to support federal data operations at the Department of Defense, Federal Aviation Administration, US Equal Employment Opportunity Commission, and others. BIRT however, just like most “commercial solutions” is not a solution that is easy to manage without the support of Actuate, and it implies high dependence on the vendor. It has the advantage of providing dashboards, and quick-ready solutions, but the dependency limits the federal agencies and departments from having their own controlled data infrastructure. Another major player in the federal data area is Salient . Salient calls their solution the Federal Mission Software Solution (FMSS). FMSS is based on agile development, it provides electronic workflows for data such as: electronic federal signatures, electronic disbursement of funds through the US treasury, electronic payment collection from citizens through www.pay.gov, and infrastructure management (among other workflows). Although some Salient solutions are deployed at the US Air Force, the US Army, Department of Commerce, and many others , Salient doesn’t provide a comprehensive workflow to federal agencies that can be followed to build a data system, rather, the models by Salient are tailored to solve specific problems. The corresponding quasi-solutions to those problems are assembled, and accordingly Salient claims that the company provides a comprehensive federal solution for data analytics. Neither BIRT nor FMSS are fully sufficient to a federal agency or department to cover many data management requirements there.
Information builders on the other hand is a tool that provides a federal solution called WebFOCUSRStat . RStat provides federal agencies with R-based advanced analytics to help with proactive decision making. Although multiple vendors provide advanced analytics through their solutions, RStat is specifically designed to federal agencies analytical needs. RStat focuses on the following fields for advanced analytics: Social Security, Tax Collection, Customs, Medicare/Medicaid, and Armed Forces. An example of a dashboard that federal users can interact with is presented in Fig. 4 below. RStat lacks descriptive and general types of analysis. It also lacks a sufficient process for data management that could be easily implemented by government workers.
Other major players in the data analytics field for government are Deloitte, Accenture, SAS, Tableau and many others. Many of these vendors would be more than keen to get into the business of building solutions to the government, however, that means that they will be in control of maintaining, updating, and improving such systems, something that the government might not be able to afford. Due to a recent study by the Association of Government Accountants (AGA) , the government hasn’t fully adapted data analytical infrastructures due to 3 major reasons: 1. A lack of budget resources (67%): getting any of the previously mentioned solutions or using any of the vendor-specific models exacerbates this problem. These vendor solutions are relatively expensive. 2. A lack of appropriate staff (53%) 3. Uncertainty as to how to develop a data analytics system (33%): deploying off the shelf solutions shall influence the agency in a negative way in terms of uncertainty . Therefore, building an in-house solution creates less ambiguity and more clarity into the future of the data system at the agency. The AGA interviewed multiple representatives from multiple Federal agencies, they categorized the agencies into 3 categories based on desire, understanding and deployment of data analytical systems: High (23% of agencies), Medium (68%) and Low level (9%). Examples of the agencies and departments interviewed and are interested in a data management model for big data analytical systems:1. US Department of Agriculture, Food and Nutrition Service (FNS): Deployed a data system called ALERT: a system for fraud detection.2. US Postal Service: Deployed a system called RADR: a system that aids in selection of areas for investigation and audit.3. Centers for Medicare and Medicaid Services: Deployed a system to identify and prevent waste, fraud and abuse in the Medicare fee-for-service program. Many other departments and agencies deployed data analytical systems, more are found under  and Fig. 5 below.
Therefore, there is a gap in literature for an in-house federal driven models that guide the implementation of a data system. Additionally, there is a lack of validation and data sharing techniques at the federal government.
That model needs to accommodate the following characteristics:
A model that facilitates data streaming between agencies.
A model that validates the data, maximizes its security and provides a standardization mechanism.
A model that is built in-house and without dependency on commercial products
A model that can handle big amounts of data
Based on what have been discussed in previous sections, and due to the immediate need for a management model that has the aforementioned four characteristics, this paper introduces: The Federal Data Management and Validation Model (FedDMV) to address all the challenges and fill the existing void. FedDMV has been used and experimented with at a federal agency. It is introduced in the next section.
The model for federal analytical systems
This section introduces the main components of FedDMV; the main contribution of this paper. The major overall challenge with big data is non-repetitive unstructured data. Structured data is much easier to manage, and if the data is repetitive, then it is also easier to predict its contents and therefore easier to control [36–38]. However, what is the best way to stream and structure data? When data resides at a federal agency, how is it validated? FedDMV provides answers to these problems.
The federal data management, streaming and validation
Besides guiding the development team through the steps of developing an analytical data system, FedDMV aims to solve three main issues:
1. Data volume challenges: Quantity of data that federal agencies import and generate is big (thus making it a big data challenge). This data is highly diverse and the speed at which data flows/updates from multiple sources is high (much of it is daily, some weekly, and some monthly/yearly).
2. Distribution challenges: The federal agency aims to build visuals and shareable solutions.
3. Data quality: Validating federal data has many challenges, and requires many resources.
FedDMV addresses these challenges and consists of the following main 4 steps (illustrated in Fig. 6):
Data Sources Variables Exploration: all the unstructured data from all sources is unified into the new FedDMV centralized database.
Variable Mapping and Columns Unification: redundant (repetitive) variables are concatenated in columns, and potential contents of the columns are identified (including their data types, size, and formats).
Data Standards Creation: this step is the main step for creating Lookup (LU) tables in the database. LUs are dictionaries for standardizing the different forms of data from multiple sources.
Data Validation: Validate and verify the data in the target Database (DB) – using an Association Rules Algorithm (discussed in the next section). Validation is done is 3 steps:
Test case collection and execution.
Test case association measurement and evaluation.
Data system refinement.
Data Security Assurance
Federal Restrictions Deployment: deploy federal regulations into the data system.
All the steps are discussed next in great detail. Figure 6 illustrates the bottom-up workflow and outcomes for FedDMV. For the first step, and as the data flows in from different sources, it goes through a knowledge-base (KB). The KB explores the variables and organize the data in a structured manner. The streaming process is shown in Fig. 7 below. The KB works as a mapping mechanism. Each variable from a source is mapped to a variable in the destination using a table to associated variables. For example, a certain commodity (auto part - tires) that is being exported from Germany to the US is represented with the following data string in one resource:
EXP TIRE, AUTO PARTS, GER, US $70.00
Similar data is presented in a different form from another data source:
Auto Tires, Germany, Trade Exports from: USA, 70 US Dollars.
Another data source would have similar data like this: GRTIRExUS70; that needs to be broken down into multiple fields and eventually migrated into data fields in the destination table. The design for the database needs to consider both (and possibly more) formats of data streams. However, similar data fall within the same data group (the data is grouped into different types/groups and each group consist of data that share common characteristics). The goal is to standardize the data format, and to have a single version of the truth data format in the FedDMV system. The dimensions that need to be standardized cover the where, when and what. That is correspondent with Geography, Time, and specific subject matter data respectively. Geography covers Countries, States, Counties, Cities, and so on. While Time addresses Years, Months, Weeks, Hours, and Minutes. Analysts then interact with the data in these tools, build dashboards, and visualizations for publication on federal websites, executive’s mobile devices and to be shared with other users or the general American public.
An increasing number of organizations have already done, or are currently planning to proceed with data streaming and migration tasks. However, most people underestimated the complexity of data streaming. According to the survey conducted by Howard and Potter, only 16% of migration projects were finished on time and on budget [36–38]. Well-defined and practically proved models and methodologies have been sought by both the research and industry communities. Morris, in his book, the author discussed streaming data from legacy systems to new destinations, from planning, organizing, to implementing data migration projects [37, 39–41]. The book was written for readers including organization executives and data migration practitioners, and it mainly presented methodologies and pitfalls. All of them are very general and from the perspective of strategies without any models or specific technical description. In the paper of Wu et al.  though, the Butterfly Methodology was proposed to solve the legacy system streaming and migration problem. It includes six major phases, each with a set of independent activities. The paper provided a comprehensive model of migration of a whole system (and mainly targeting the data of a system), instead of focusing on data migration only. The authors also stated that they need more practical experiments to explore the relationships between system variables. Furthermore, the papers of Haller, Thalheim and Wang focused on the theoretical perspective of data migration [43, 44]. The authors investigated the connection between two types of models: abstract model and concrete model. Then, they re-defined the Extract-Transform-Load (ETL) process in data streaming, and discussed two types: property-preserving transformation and property-enhancing transformation. Also from the theoretical perspective, Spivak considered data migration with mathematical concepts . In the book by Anavi-Chaput et al. , the authors discussed the process of migrating PeopleSoft Applications from Oracle to DB2 [46, 47] with highly detailed steps and scripts. These services include Transportable Tablespaces (TTS), Trasportable Databases (TDB), Data Pump, and Recovery Manager (RMAN). Validation in migration is a very important aspect as well, studies looked at migration from a testing/validation perspective [48–50]. Similarly, several other online articles written by Scheier , Burry , Klazema , and Levine  discussed the average process and common errors of data migration, but none of the methods injected validation into streaming or paid attention to its risks and increased importance. Moreover, how can analysts trust this streamed data if it is not validated? FedDMV uses a method called ART (Association Driven Testing) or ADT (Association Driven Testing) - both ART and ADT are used interchangeably - to validate the streamed data system, and find potential errors. A sample validation SQL code is in Fig. 8 below, and streaming SQL sample in Fig. 9. However, in big data, finding an error is difficult, due to size and variety of data; so how does FedDMV determine the location of these errors? The next section introduces FedDMV’s validation and addresses all the mentioned challenges.
Federal data validation
As it is well known, if the data migrated is not sufficiently validated, it would be extremely risky to the owners of the data in the future. Data validation is not used only to compare data between the origin and the destination, but also to ensure the semantical correctness, completeness, consistency, and the interoperability . Especially, nowadays, simple databases have been growing tremendously into “big data”, many business intelligence and decision making highly rely on “data”, lower data quality may result in unsuccessful business strategies, which will further bring severe financial loses to any organization. Therefore, migrating data is always risky, and should be planned, implemented, and validated seriously. To mitigate the risks of the data anomalies, the research community proposed different state-of-the-art models and solutions. Researchers looked into common errors and risks of data migration that should be taken into consideration, but didn’t provide any solutions [46 ,49, 50, 52–55]. Different organizations may have different requirements about how to evaluate the data quality, and there is not a single universal technique that can solve it. Woodall et al. for example, proposed a hybrid approach of multiple methods to dynamically assess the data quality [56, 57]. However, based on the recent study by NIST [58, 59], the data error is not the challenge, it is actually the ability to define the locations of errors. That is listed as the most time-consuming activity of data validation. In the study, the NIST researchers compiled a vast number of software and data analytics projects and reached the following conclusion:
“If the location of bugs can be made more precise, both the calendar time and resource requirements of testing can be reduced. Modern data and software products typically contain millions of lines of code. Precisely locating the source of bugs in that code can be very resource consuming.”
Literature therefore, has a gap in error allocation; no or few AI-driven method was introduced to help analysts and engineers pin-point to errors that would affect the overall health of a data system. Some methods touched on applying AI logic to software engineering, and some others developed means to locate errors that are not based on a solid AI approach. FedDMV uses association rules for that purpose, ART is presented next.
Association rules validation and testing
Association Rules (AR) is one of the most commonplace data analytical models . AR are intended to identify patterns of the type: “Action B often comes after action A, and is followed by Action C”. One of the more well-studied problems in data mining is the search for AR in market basket data, or mostly referred to Market Basket Analysis (MBA). MBA allows businesses to realize which commodities are bought together, and therefore, they put them close to each other on store shelves. The outcome of the AR model is a set of rules. These rules include consequents (B → C), antecedents (A → B), support, and confidence of the rules. Confidence measures the reliability of the inference made by a rule. For a given rule B → C, the higher the confidence, the more likely it is for C to be present in transactions that contain B. Confidence also provides an estimate of the conditional probability of C given B.
Support is an important measure as well, because a rule that has very low support may occur simply by chance. Association Rules outputs should be interpreted with thoughtfulness. The inference made by an association rule does not necessarily imply causality. Instead, it suggests a strong co-occurrence relationship between items in the antecedent and consequent of the rule.
Support and confidence are bound with these rules:
In this paper, ART aids in locating errors in data after streaming, by looking at the consequents of existing data errors . For example, a data error occurred in module 1 function 2 (M1F2) of streaming; due to the module’s importance, this module effects other modules in the system. Due to the AR statistical model, M7F10, M10F11 are the most two functions that historically effect errors in M1F2 (M7F10, M10F11 → M1F2). While M6F1 is the function that will be effected due to the error that exist in M1F2. This M1F2 → M6F1 rule is what the engineer is looking for. However, they need to confirm that the rule has high confidence and acceptable support, they look at the confidence, and realize that its 0.805, which is considered very high (a confidence of 0.75 and up is considered very high, 0.50–0.75 is high, 0.30–0.50 is low, 0.30 and lower is very low). After the current testing iteration, the engineering team should perform heavy testing on the location that the ART engine recommended: M6F1. The testing needs to be done before the actual release of the data to the federal employee.
Data Collection: In order to increase the accuracy of any data mining model, a fair amount of high quality data required. More data helps the analysts and the engineers get a better quality model that could be more insightful.
Model Development and Data Training: In this stage, the AR model is built and the data is trained.
AR Model Outputs: Outputs of the model are antecedents, consequents, confidence and support.
Sort all Predictions: Consequent predictions are sorted by confidence, top predictions with highest confidence results are then considered for testing with data streaming.
Federal intelligent testing: using the outcomes from step 3 and the predictions from step 4, testing is performed and focused on the system modules present in the AR model’s consequents.
Federal data security
Database security is about preventing malicious users from accessing or modifying data stored in databases. Similar to the generic security concept, three key aspects determine how secured a database is: Confidentiality, Integrity, and Availability. Confidentiality refers to only disclosing data to authorized users. To verify a user’s identity and control access to data, database management systems (DBMS) use different methods of authentication [61–63]. Integrity refers to protecting the database from unauthorized writing. Data stored in databases should not be modified improperly, (i.e. corrupted). At the age of “big data”, protecting database is one of the critical missions. Access control was one of the earliest database security measures proposed and widely used. In most cases, access control models can be categorized into three classes: discretionary access control (DAC), mandatory access control (MAC), and role-based access control (RBAC) [62–65] – this is the method adopted by FedDMV. The reason for that is based on the notion that federal data control is dependent on employees’ roles, access levels, and aggregations. The next section introduces the experimental work that was performed with FedDMV.
FedDMV experimental studies
This section introduces the main three experimental works of this paper, the first is through comparing a big data analytics system built using FedDMV to another built through traditional relational database; the second by surveying feedback on FedDMV’s tool from federal employees and analysts. The third is a test of ART, FedDMV’s validation method. The subsections present pros and cons of the two compared processes, outcomes of the comparison, results of the survey, ART and the lessons learnt.
Setup for the experiments
FedDMV was deployed to develop a data system at a US federal agency. The steps presented in the previous section are applied to a new data system development process. This section compares FedDMV to the traditional process that most federal agencies are currently using. A big data system was built using FedDMV, and another one (with similar size) was built through Traditional Relational Database approaches (that is referred to as TRDb here). The two process’s steps are presented in Table 2. The federal agency’s identity and information are not presented in this paper due to government data restrictions, however, the project that is managed by George Mason University is funded by the United States Department of Agriculture (USDA). For further details, please consult with the authors.
Data was collected to run the experiment a total of 633 rows was assembled. The system under test has 11 modules, with an average of 58 functions. Due to proprietary reasons, the nature and details of the system can’t be exposed.
The FedDMV process was performed to build a federal data system at USDA, at each stage, certain decisions were taken. The code shown in Fig. 9 creates a cursor that goes through all data and migrates it to tables in a destination database. Such procedures could be processed using SQL engines. The federal employees require friendly graphical user interfaces to be able to interact with the data. That includes the ability to upload, update, delete, and retrieve data from the FedDMV system. A tool is part of the process for data management. Developing a tool goes through all the steps of software development, coding, testing, and deployment. An example screenshot of the tool is shown in Fig. 10 (the tool provides a GUI for analysts to interact, stream and validate the data).
As for ART, a sample data that is used as part of the experiments is shown in Fig. 11 (after sorting rows by confidence). The figure shows the errors that previously occurred in Column A, the Antecedents in Column B, the Consequents in Column C (which what the engineer look for at each stage). The figure also shows Confidence, Lift and the Support in columns D, E, and F. This data was collected and generated using ART throughout the development and testing process. The results of the three experiments are presented next.
Experimental results and conclusions
This section presents the experimental results of the three studies, pros and cons of FedDMV and TRDb, conclusions and future work.
Time consumption and lessons learnt were collected from both processes, time consumption for FedDMV is significantly less than that of TRDb (45 vs. 81 months). Results for FedDMV are shown in Table 3 below, results for TRDb are in Table 4. Furthermore, advantages of FedDMV and disadvantages of TRDb are in Table 5.
For data management models, it is very important to include the end-users and collect their feedback. Therefore, we presented the FedDMV system to 350 governmental employees and analysts and collected their feedback on FedDMV’s usability. The feedback was mostly positive. Feedback was categorized into the following five groups: FedDMV’s tool is user friendly, FedDMV is an inevitable change (willing to accept it), FedDMV is easy to follow, haven’t used the tool, dislike FedDMV and its tool. See results illustrated in Fig. 12. We noticed that new federal hires are usually early adopters of such data technologies, and are excited to see government systems move in the right direction and adopt new technologies. Nonetheless, many employees whom been with a federal agency for 15 years or more are very strict in terms of resisting change and refusing to change their ‘old’ ways.
For example, 9% of the analysts gave us negative feedback that us relevant to the notion that they don’t see the need to move away from more manual processes. However, 71% gave great or good feedback, and 20% thought that this change is necessary and eventually inevitable.
As for ART, the results were collected after performing 11 iterations of streaming and testing. The results of the experiment are as follows: the riskiest module across all iterations is M6F1 (more than 150 occurrences); Refer to Fig. 13. This module is expected to have the highest number of errors. However, the module that the model has highest confidence that it will have errors is M8F14 (the only module with a confidence of 1). Refer to Fig. 14. As it is evident from the experimental results, many functions that are directly affected by errors in the data system are the first function in the module. As Fig. 13 suggests, M6F1, M3F1, M5F1 and M7F1 are modules with evident high risk. The AR model’s average confidence of the data system is 0.55011; therefore, most recommendations in this model are dependable as testing candidates. The data system module used for the experiment was released on time, and on budget and is currently in use by federal team. We aim to provide more experiments to ART in the future. Conclusions and future work are presented next.
Discussion and Conclusion
The 4-step model in this paper aims to guide the deployment process of a federal big data analytical system. For any data model, the following four factors are used to evaluate its efficiency and effectiveness:
Resource consumption rates: is the process expensive? How many hours? Manpower?
Complexity: how difficult is the model to follow?
Practicality: is the model merely theoretical or is it a good fit for real world projects?
Validity: ensuring that the data and the processes are valid
In the experimental study of this paper, FedDMV is compared with TRDb, a traditional process for building data systems. To assess the four points: in terms of resource consumption, FedDMV presented a significant improvement over traditional federal processes (45 months vs. 81 months), and was successful in delivering a big data system at a federal agency in a timely manner. In terms of complexity and practicality, FedDMV has 4 clear steps, and after working with multiple engineers at the federal agency, there is a general consent that FedDMV is easy to follow (the tool survey shows that as well). As for validity, FedDMV focuses of validation through ART.
Future work of our research includes the following:
Perform more experiments for ART and general quality assurance of data systems.
Deploy big data systems with other federal agencies. That will put FedDMV to test with different types of data, and different types of processes.
Deploy FedDMV with smaller types of systems and evaluate its feasibility with that.
Compare FedDMV to more data analytics models, in terms of time consumption, usability and applicability - such as PACER and BIRT.
Provide more software tools for managing the FedDMV process, and aid in project planning, tracking resources, and monitoring the 4 steps of the process.
FedDMV focuses on major federal big data aspects such as validation, streaming, security, and automation. In this age of big data, FedDMV is introduced as an effective and efficient 4-step model to follow, and is a very strong candidate for federal agencies that aim to develop new data systems.
Affordable Care Act
Association of Government Accountants
Database Management Systems
Federal Data Management and Validation
Federal Mission Software Solution
Government Accountability Office
Market and Trade Economics Division
National Agricultural Statistics Service
Public Courts Online Access System
Traditional Relational Database
United States Department of Agriculture
Turing AM. Computing machinery and intelligence. Mind. 1950;49:433–60.
McCarthy J. Programs with common sense. In: Proceedings of the Symposium on Mechanisation of Thought Processes, vol. 1. London: Her Majesty’s Stationery Office; 1958. p. 77–84.
Koh HC, Gerald T. Data Mining Applications in Healthcare. J Healthc Inf Manage. 19. 2.
Jing L. Data Mining Applications in Higher Education, SPSS Executive Report. 2004.
Book: Industry Applications of Data Mining. Chapter 8, 1999.
Harmann M. The Role of Artificial Intelligence in Software Engineering. Published report at the Crest Center, University College London.
An Immediate Release by the Office of Science and Technology Policy. Executive Office of the President. The Big Data Initiative. 2012.
The White House Big Government Initiative: http://www.whitehouse.gov/Open/.
White Paper by Accenture. Accenture Federal Services, Federal Analytics and Big Data. 2012.
Letter from James Madison to W.T. Barry (Aug. 4, 1822), reprinted in The Writings of James Madison (Gaillard Hunted.)
A White Paper by CC Pace Systems. Agile in the Federal Government. 2014.
Batarseh FA. Incremental Lifecycle Validation of Knowledge-Based Systems through CommonKADS. PhD Dissertation Registered at the University of Central Florida and the Library of Congress; 2012.
DePillis L. The Way Government does Tech is Outdated and Risky. A Report Published at the Washington Post; 2013.
Smith M, Cohen T, A CNN Report on the ACA Website. Problems with Health Website. 2013.
United States Government Accountability Office (GAO). Software Development - Effective Practices ad Federal Challenges in Applying Agile Methods, GAO-12-681, gao.gov. 2012.
Batarseh F, Gonzalez A. Predicting Failures in Contextual Software Development through Data Analytics. Proc Springer Softw Qual J. 2015.
Ming-Tun Yu H. Designing Software To Shape Open Government Policy. A Dissertation Presented to the Faculty of Princeton University in Candidacy for the Degree of Doctor of Philosophy. The Department of Computer Science, Princeton University; 2012.
Hahn R. Government Policy toward Open Source Software: An Overview. Chapter 1. Published at the Brookings Institute.
National Agricultural Statistics Service: http://quickstats.nass.usda.gov/.
City of Chicago Data Portal: http://data.cityofchicago.org.
San Francisco Governmental Data Website: http://data.sfgov.org.
Swish-Data - Data Performance Architects, Hadoop Uses Cases: Big Data for the Government, http://www.swishdata.com/index.php/blog/article/hadoop-use-cases-big-data-for-the-government.
MapR: Big Data and Apache Hadoop for Government: http://www.mapr.com/solutions/industry/big-data-and-apache-hadoop-government.
Horton Works by Apache: http://hortonworks.com/industry/government/.
Microsoft’s SQL Server: http://www.microsoft.com/SQLserver.
Batarseh F, Gonzalez AJ. Incremental Lifecycle Validation of Knowledge-Based Systems through CommonKADS. Published at the IEEE Transactions on Systems, Man and Cybernetics (TSMC-A); 2012.
Batarseh F. Context-Driven Testing. In: Context in Computing: A Cross-Disciplinary Approach for Modeling the Real World. Patrick Brezillon, Editor (Sorbonne University- Paris VI). Springer Verlag; 2015. ISBN: 978-1-4939-1886-7.
Scacchi W. Process Models in Software Engineering. Institute for Software Research, Encyclopedia of Software Engineering. 2nd Ed. John Wiley and Sons; 2001.
Gartner’s Magic Quadrant: http://www.gartner.com.
A Report by Actuate. Reporting and Data Analytics for Federal Applications. Actuate BIRT; 2014
Report S. Federal Mission Software Solution. 2014.
Information-builders. Predictive Analytics for Federal Government. New York: Product Brochure; 2011.
Corporate Partner Advisory Group, Research Series. Leveraging Data Analytics in Federal Organizations, Report no. 30. 2012.
Howard P, Potter C. Data Migration in the Global 2000 - Research, Forecasts and Survey Results. 2007. p. 29.
Morris J. Practical Data Migration. 3rd ed. Swindon: British Informatics Society Ltd; 2006.
Inmon W, Linstedt D. Data Architecture: A Primer for the Data Scientist: Big Data, Data Warehouse and Data Vault. Morgan Kaufmann; 2014.
A Report by IBM. The Hidden Costs of Data Migration - Strategies for Reducing Costs and Quickly Achieving Return on Investment. 2007.
Wei B, Chen T. Verifying Data Migration Correctness: The Checksum Principle. RTI Press; 2014. OP-0019-1403.
Department of Education, Office of Federal Student Aid. Data Migration Roadmap: A Best Practice Summary, Version 1.0. 2007.
Wu B, Lawless D, Bisbal J, Richardson R, Grimson J, Wade V, Sullivan D. The Butterfly Methodology: A Gateway-free Approach for Migrating Legacy Information Systems, Proceedings of the 3rd IEEE International Conference on Engineering of Complex Computer Systems (ICECCS97). Como: Institute of Electrical and Electronics Engineers; 1997. p. 200–5.
Haller K, Matthes F, Schulz C. Testing & Quality Assurance in Data Migration Projects, Proceedings of the27th IEEE International Conference on Software Maintenance ICSM. 2011.
Thalheim B, Wang Q. Data Migration: A Theoretical Perspective. Data Knowl Eng. 2013;87:260–78.
Spivak D. Functional data migration. Inf Computing. 2012;217:31–51.
Oracle’s Document: Move to Oracle Database with Oracle SQL Developer Migrations.
Oracle White Paper. Migrating Oracle Databases. 2014.
Haller K, Matthes F, Schulz C. Testing & Quality Assurance in Data Migration Projects. In: 27th IEEE International Conference on Software Maintenance ICSM. 2011.
Fine L, Keogh B, Cretin S, Orlando M, Gould M. How to Evaluate and Improve the Quality and Credibility of an Outcomes Database: Validation and Feedback Study on the UK Cardiac Surgery Experience. BMJ. 2003;326(7379):25–8.
Flamos A, Doukas H, Psarras J. Data Validation Platform for the Sophisticated Monitoring and Communication of the Energy Technology Sector. Renew Energy. 2010;35(5):931–5.
Scheier R. Data Migration Strategies and Best Practices. An Article Published at TechTarget.
Burry C, Mancusi D. How to Plan for Data Migration, Proceedings of the Advanced Information Systems Engineering: 21st International Conference. 2004.
Klazema,A. Data Migration Strategies for Businesses. http://blog.udemy.com/data-migration-strategies/. An Udemy Blog.
Levine R,A web-report by Alta Flux Corporation. Data Migration Strategies. 2013.
Anavi-Chaput V, Arrell K, Baisden J, Corrihons R, Fallon D, Siegmund L, Sokolof N. Planning for a Migration of PeopleSoft 7.5 from Oracle/UNIX to DB2 for OS/390. Poughkeepsie; 2000. p. 148.
Woodall P, Borek A, Parlikad A. Data quality assessment: The Hybrid Approach. Info Manag. 2013;50(7):369–82.
A Report by IBM. Best practices for data migration - Methodologies for assessing, planning, moving and validating data migration. Somers; 2009. p. 16.
National Institute of Standards and Technology: WWW.NIST.GOV.
Planning Report for NIST (US Department of Commerce). The Economic Impacts of Inadequate Infrastructure for Software Testing. 2002.
Association Analysis: Basic Concepts and Algorithms, Introduction to Data Mining, Chapter 6, pp. 327–414, by: Pang-Ning tan, Michael Steinbach, and Vipinkumar.
Oracle’s Database Concepts Guide: From the Oracle Help Center.
De Capitani di Vimercati S, Foresti S, Samarati P. Recent Advances in Access Control. In: Handbook of Database Security. 2008; pp. 1–26.
Bertino E, Jajodia S, Samarati P. Database security: Research and practice. Information Systems. 1995.
Hore B, Mehrotra S, Hacigümüç H. Managing and Querying Encrypted Data. In: Handbook of Database Security. 2008. p. 1–26.
Oracle’s Database Advanced Security Administrator’s Guide: From the Oracle Help Center.
The authors would like to convey their thanks to the staff of the United Stated Department of Agriculture - Economic Research Services (ERS); especially the Applications Development Branch (ADB) and the Market and Trade Economics Division (MTED). The views and opinions expressed in this research paper are those of the authors and do not reflect the official policy or position of any agency or department of the U.S. government.
Additionally, an acknowledgement goes to the following George Mason University students for their help and hard work with the federal projects: Gowtham Ramamoorthy, Manish Dashora and Samantha Dcosta.
Availability of data and materials
The actual Data is not shared because this is federal data and parts of the data set are private or confidential, therefore, we are not allowed to share it. However, to get a sample data set visit: http://ers.usda.gov/data-products.aspx. All the software tools used in the experiments and as part of FedDMV are available via the online sharing website (Dropbox), upon request from the author.
FB worked on the overall design and development of the FedDMV model and the associated system, executed the experiments, collected data for the comparison study, and interacted with the federal employees on frequent basis. FB wrote multiple sections of this manuscript and provided the illustrations. RY worked with the federal agency management to setup the experiments. RY wrote multiple sections of the paper and provided multiple insights. LD developed the tool, and executed some of the experiments. LD also wrote some parts of the paper. All authors read and approved the final manuscript.
The authors declare that they have no competing interests.
Consent for publication
Ethics approval and consent to participate
About this article
Cite this article
Batarseh, F.A., Yang, R. & Deng, L. A comprehensive model for management and validation of federal big data analytical systems. Big Data Anal 2, 2 (2017). https://doi.org/10.1186/s41044-016-0017-x