Thank you for Subscribing to CIO Applications Weekly Brief

Revitalizing Data Collection with Next- Gen Technology
Marci Augustine, Director, Business Intelligence & Analytics, Xo Group Inc.


Marci Augustine, Director, Business Intelligence & Analytics, Xo Group Inc.
Digital data collection utilized to analyze user behavior provides seemingly endless opportunity to know every imaginable detail. In the early days the possibilities were overwhelming and often challenging to capture, thus more care was taken to ensure the most critical data points were captured properly. The hallmark of a good analysis is that it not only answers the questions being asked of it, but it also inspires new ones. So naturally, as analytical capabilities and output increase the next questions, such as: “How can we get this information for more business processes?”, “When can we have this for all processes?” The natural next step is to find a better way to get more data faster.
As data collection systems and tag managers are developed, organizations are able to leverage a software infrastructure that makes it easy to capture any data points with fairly simple JavaScript code changes and easy-to-use configuration tools that don’t require (much) intervention from software engineers.
The hallmark of a good analysis is that it not only answers the questions being asked of it, but it also inspires new ones, so naturally, as analytical capabilities and output increase the next questions
Sorting through all of these questions and reconciling data is time-consuming and frustrating. Product and business owners have the time to think through their data needs in detail-they need to be focused on the resulting decisions. At the same time, data teams are not typically staffed to handle detailed curation for each individual stakeholder. These groups have to work together to develop a system to facilitate quality data collection at the speed of business. This can best be accomplished with a two-pronged approach, implementing technical safeguards to back up a well-managed process for identifying data requirements.
The key to getting support for this kind of approach is to first make it easy for people to start using and then demonstrate immediate value by showing results. Successful rollouts will start with a handful of basic templates that handle most use cases in the organization with close collaboration between business and technical teams to ensure everyone is on board. By templatizing the data requirements where possible, less time is required to identify what needs to be captured up front. There will be cases that don’t fit into the predefined templates, but those will be the exceptions, not the rule. Leveraging these best practices and evangelizing them across the organization allows data teams to create more meaningful, clean data without introducing huge bottlenecks in the process. At the end of the day, humans make mistakes, even with the best processes outlined and executed. That’s where a technical solution that can provide validation can provide a last line of defense to ensure the cleanest data possible for our most important business processes. A tool, whether custom built or off the shelf, which allows for the configuration of validation rules for each type of data collected, as well an opportunity to correct bad data before it goes too far. The robustness required in this tool will vary depending on the organization and the volume data collected, but it is a critical piece of the data quality puzzle.
Demonstrating success with a program like this among a pilot group will facilitate broader adoption in the organization. As working groups see analysts leveraging this process spending fewer cycles reconciling data or tracking down the meaning of specific data points and spending more cycles providing insights, they are incentivized to make the investment themselves. This can then have a spiraling effect across the organization with groups making the time to prioritize adopting these tools and process and over time clean, quality data readily available will be the rule, not the exception.
I agree We use cookies on this website to enhance your user experience. By clicking any link on this page you are giving your consent for us to set cookies. More info
Featured Vendors
-
Jason Vogel, Senior Director of Product Strategy & Development, Silver Wealth Technologies
James Brown, CEO, Smart Communications
Deepak Dube, Founder and CEO, Datanomers
Tory Hazard, CEO, Institutional Cash Distributors
Jean Jacques Borno, CFP®, Founder & CEO, 1787fp
-
Andrew Rudd, CEO, Advisor Software
Douglas Jones, Vice President Operations, NETSOL Technologies
Matt McCormick, CEO, AddOn Networks
Jeff Peters, President, and Co-Founder, Focalized Networks
Tom Jordan, VP, Financial Software Solutions, Digital Check Corp
Tracey Dunlap, Chief Experience Officer, Zenmonics