Addressing Big Data One Byte At A Time
Strapline: Industry Warehouse
![rene-keller-informationmosaic rene-keller-informationmosaic](/sites/default/files/styles/landscape_750_463/public/import/IMG/995/216995/rene-keller-informationmosaic-580x358.jpg.webp?itok=ijYwaywF)
Unlike other recent industry buzzwords that have trended but in actuality turned out to be “much ado about nothing,” the term “big data” represents important and real challenges to the technology and operations departments of financial firms. The capability to centrally maintain large volumes of data from disparate sources, and process that data into useful information will require a unique set of advanced tools. As the complexity, as well as the sheer volume, of the data grows, solutions will have to be dynamic and flexible in their design.
The first challenge is consolidating data—whether it is reference data, market data, corporate action data, or even pricing—from several sources. While financial firms are reducing the number of relationships they maintain with data providers, numerous feeds currently exist across the enterprise and pose a real problem today. In addition, there were actual buying decisions behind each of these relationships, and a dependence upon this information, somewhere in the enterprise, has been created. The reduction of feeds may therefore be slow to materialize.
The solution obviously requires scalable architecture. Perhaps more important is the need to deploy flexible mapping utilities that can find, and even create, common fields across different file layouts. A lack of a common industry standard, at present, means various files have to be cleansed and normalized against a defined format. The resultant layout will need to be adaptable to future changes. This requires a solution with a flexible data model.
Once a consolidated format is agreed, a presentation layer needs to be designed and implemented so the information can be of value to the end-user and client. As a result of recent market downturns, there is a growing demand throughout the industry, from regulators and investors alike, for quality data; “assurance” and “transparency” being two more important and very real terms. To address this, the solution requires sophisticated rules engines coupled with the automated mapping techniques mentioned above.
In conjunction with the presentation layer, there is an acknowledged demand in the marketplace for business intelligence tools to address big data. Detailed analysis is the next logical step in processing the data and turning it into useful information. Again, this requires a scalable and high-end processing application, as well as a sophisticated analytics solution. Besides the typical day-to-day requirements of any analytics solution, such as report and inquiry creation, a cutting-edge analytics engine should be able to generate output such as “what-if” scenarios and “in/out of the money” alerts. To make this functionality worthwhile, the solution needs to be online and in real-time. If the generation of new reports, inquiries, and alerts could be done by an end-user as opposed to having them coded by the IT department, valuable technical resources would be freed up and time to rollout would be drastically reduced.
Lastly, as a result of recent economic declines, a very real challenge comes from the need for global regulation. While all market jurisdictions generally agree on what is needed for increased transparency and systemic risk mitigation, the sources of this information are as segmented as the markets themselves. Therefore, the need for central processing and consolidation on a global basis is crucial. A pressing case-in-point is the introduction and global endorsement of the legal entity identifier (LEI), created to indicate the name, location, electronic address, and legal status of an organizational entity. The LEI in itself requires the flexible data model mentioned earlier, as well as modern mapping tools. Not only are financial firms required to implement the LEI in current data layouts, but they must also link to historical transactions generated pre-LEI. Data model flexibility and openness are instrumental to incorporating future global regulatory requirements in a timely and accurate way.
In closing, the challenges posed by big data are significant and exist right now. However, with the implementation of modern, robust solutions, not only can the challenges be met but opportunities for value-added features for firms and their clients can be created.
Only users who have a paid subscription or are part of a corporate subscription are able to print or copy content.
To access these options, along with all other subscription benefits, please contact info@waterstechnology.com or view our subscription options here: http://subscriptions.waterstechnology.com/subscribe
You are currently unable to print this content. Please contact info@waterstechnology.com to find out more.
You are currently unable to copy this content. Please contact info@waterstechnology.com to find out more.
Copyright Infopro Digital Limited. All rights reserved.
As outlined in our terms and conditions, https://www.infopro-digital.com/terms-and-conditions/subscriptions/ (point 2.4), printing is limited to a single copy.
If you would like to purchase additional rights please email info@waterstechnology.com
Copyright Infopro Digital Limited. All rights reserved.
You may share this content using our article tools. As outlined in our terms and conditions, https://www.infopro-digital.com/terms-and-conditions/subscriptions/ (clause 2.4), an Authorised User may only make one copy of the materials for their own personal use. You must also comply with the restrictions in clause 2.5.
If you would like to purchase additional rights please email info@waterstechnology.com
More on Data Management
S&P sees strong demand for GenAI tools as leadership changes hands
The data provider released several AI-enabled tools and augmentations to existing platforms in 2024 and plans to continue to capitalize on the technology in 2025.
To modernize loan markets, making data more accessible is key
Wilmington Trust is using AccessFintech’s Synergy platform to ditch faxes and emails in the increasingly popular asset class.
Lucrative market data deal with LSEG fuels Tradeweb’s record quarter
The fixed-income trading venue realized gains from its 2023 deal with the London Stock Exchange Group, amid soaring revenues from market data providers industry-wide.
Is overnight equities trading a fad or the future?
Competition is heating up in US equity markets as more venues look to provide trading from twilight to dawn. But overnight trading has skeptics, and there are technical considerations to address.
DTCC revamps data distribution, collection efforts with cloud, AI
The US clearinghouse is evaluating the possibilities that cloud and AI offer to streamline the processes by which it collects and makes data available to market participants.
European exchange data prices surge, new study shows
The report analyzed market data prices and fee structures from 2017 to 2024 and found that fee schedules have increased exponentially. Several exchanges say the findings are misleading.
We’re running out of datacenters! (But maybe AI can help?)
The IMD Wrap: Datacenter and cloud adoption is being pushed to its limits by AI. Will we simply run out of space and power building AIs before AI figures out how to fix it?
FCA: Consolidated tape for UK equities won’t happen until 2028
At an event last week, the FCA proposed a new timeline for the CT, which received pushback from participants, according to sources.