Big-Time Data Terminology
The term "big data" is so broad that when commenting about data management and the industry, it's better to consider topics such as data quality, data consistency or deriving value from data – or at least discuss matters in those terms.
A presentation given this past week by Pierre Feligioni, head of real-time data strategy at S&P Capital IQ, defined "big data" as "actionable data," and sought to portray big data concerns as really being about four issues: integration, technology, content and scalability.
Integration, particularly the centralization of reference data, is the biggest challenge for managing big data, as Feligioni sees it. While structured data is already quite "normalized," unstructured data, which can include messaging, emails, blogs and Twitter feeds, needs to be normalized.
Unstructured data is fueling rapid exponential growth in data volumes, justifying the name "big data." Data volumes are counted in terabytes (1,000 gigabytes), or even petabytes (1,000 terabytes). When it comes to unstructured data at those levels, central repositories that can collect and normalize data – and coordinate it with structured data – are a must, Feligioni contends.
Technology and scalability the building blocks necessary to make such central repositories functional, as he describes it. Natural language processing and semantic data approaches are also being applied. "The biggest challenge is understanding documents and creating analytics on top of this content, for the capability to make a decision to buy or sell," says Feligioni.
Scalability makes it possible to process more and more information, and is achieved through new resources, such as cloud computing, which carry their own issues and require additional decisions [as described in my column two weeks ago, "Cloud Choices"].
Everything that Feligioni calls part of "big data" actually revolves around getting higher quality data by incorporating more sources and checking them against each other to keep that data consistent. It's also about creating new value from data that can be acted upon by trading and investment operations professionals.
So, whatever buzzwords one uses, whether "big data" or sub-categories under that umbrella, what they are really talking about is quality, consistency and value. Other terms just describe the means.
Only users who have a paid subscription or are part of a corporate subscription are able to print or copy content.
To access these options, along with all other subscription benefits, please contact info@waterstechnology.com or view our subscription options here: http://subscriptions.waterstechnology.com/subscribe
You are currently unable to print this content. Please contact info@waterstechnology.com to find out more.
You are currently unable to copy this content. Please contact info@waterstechnology.com to find out more.
Copyright Infopro Digital Limited. All rights reserved.
You may share this content using our article tools. Printing this content is for the sole use of the Authorised User (named subscriber), as outlined in our terms and conditions - https://www.infopro-insight.com/terms-conditions/insight-subscriptions/
If you would like to purchase additional rights please email info@waterstechnology.com
Copyright Infopro Digital Limited. All rights reserved.
You may share this content using our article tools. Copying this content is for the sole use of the Authorised User (named subscriber), as outlined in our terms and conditions - https://www.infopro-insight.com/terms-conditions/insight-subscriptions/
If you would like to purchase additional rights please email info@waterstechnology.com
More on Data Management
After contentious Opra upgrades, vendors brace for a faster future
Upgrades to the datafeed widely used to gauge the current market price for options contracts went into effect in February after three separate delays, which market participants say were caused by persistent bandwidth issues at some important recipients.
The IMD Wrap: No more turf wars, or why CDOs should heed the Voice of the CTO
Max reviews how our recent Voice of the CTO series has implications for those beyond a firm’s technology function, and how communication and collaboration between tech, data, and leadership will deliver better results.
Dark horse: Deutsche Börse building dark pool
New functionality allowing exchange members to execute sweep trades comes hot on the heels of European rival Euronext launching its own dark pool.
Man Group’s proprietary data platform is a timesaver for quants
The investment firm’s head of data delves into its alt data strategy and use of AI tools to boost quant efficiency.
Waters Wrap: The tough climb for startups
Anthony speaks with two seasoned technologists to better understand why startups have such a tough time getting banks and asset managers to sign on the dotted line.
As crypto ETFs become reality, benchmark providers take center stage
The SEC’s approval of the first spot bitcoin ETFs will expose a growing number of traditional market participants to the maturing world of crypto data, a moment that some—such as CF Benchmarks, BlackRock’s benchmark provider—have been eagerly awaiting.
The IMD Wrap: Blowing tires, engines and budgets in F1 and financial data
There are many similarities between Formula One and financial data—except when it comes to how much you can spend and how much spend contributes to success.
This Week: Overbond, Northern Trust, FIS/Torstone, Trading Technologies, and more
A summary of the latest financial technology news.
Most read
- Women in Technology & Data Awards 2024: All the winners
- Man Group’s proprietary data platform is a timesaver for quants
- Dark horse: Deutsche Börse building dark pool