By Calum MacLeod, EMEA Director at Venafi
Organisations of all sizes and industries maintain extensive financial, customer and mission-critical business data. However, when sensitive information is misused or compromised, organisations will often pay a heavy price. Recent high-profile security breaches have cost millions in revenue and lost opportunities. These fears, along with new security standards and regulations, have driven IT professionals to deploy encryption more broadly.
The problem is that, having done so, the encryption keys used to secure data become the figurative “keys to the kingdom.” The key (and not the data itself) becomes the entity that must be safeguarded. Efforts to manage these keys manually, however, represent a significant security risk and become operationally challenging, especially as encryption is deployed across disparate systems and applications.
Organisations are struggling to properly manage and control these rapidly multiplying certificates and keys to prevent security breaches, system downtime and other disasters. It’s a catch 22 situation – but it doesn’t have to be.
Before we can solve the problem of enterprise key and certificate management (EKCM), we must first fully understand the challenges faced:
- Certificates that are not renewed and replaced before they expire can cause serious downtime and outages
- Regulations and requirements (like PCI-DSS) demand stringent security and management of cryptographic keys and auditors are increasingly reviewing the management controls and processes in use
- Private keys used with certificates must be kept secure or unauthorised individuals can intercept confidential communications or gain unauthorised access to critical systems. Failure to ensure proper segregation of duties means that admins who generate the encryption keys can use them to access sensitive, regulated data
- The average certificate and private key require four hours per year to manage, taking administrators away from more important tasks and costs hundreds of thousands of pounds per year for many organisations
- The rollout of new projects and business applications is hindered because of the inability to deploy and manage encryption to support the security requirements of those projects
- If a certificate authority is compromised or an encryption algorithm is broken, organisations must be prepared to replace all of their certificates and keys in a matter of hours
The simple fact is that certificates and private keys play a critical role in securing data and systems across all types of organisations. Having understood the risks of unmanaged encryption deployments, it is imperative to utilise EKCM best practices.
EKCM Best Practice
The effective management of certificates and private keys involves multiple individuals and groups. It is critical to establish clear and concise responsibilities for the various stakeholders. This helps ensure that nothing gets overlooked and multiple parties aren’t duplicating work to other projects.
The critical starting point in any certificate and private key management strategy is to create a comprehensive inventory of all certificates, their locations and responsible parties. This is not a trivial matter because certificates are deployed in a variety of locations by different individuals and teams – it’s simply not possible to rely on a list from a certificate authority. Adhering to the below practices will ensure that no certificates are missed:
Import from Certificate Authorities
Gather what you already know about the certificates from existing certificate authorities. It is very dangerous to assume that an import from your known CAs will provide an accurate inventory of all certificates; it’s merely a starting point that must be augmented by discovery.
Individual Import from Admins
Network and agent-based discoveries can take time and it may not be possible to perform them in all corporate locations. That makes it critical to educate administrators and make sure they are proactively reporting any certificates they are aware of and adding them to the inventory.
Perform Network Discovery
Perform a network discovery to find certificates that are present on a listening port such as HTTPS. Start by gathering your network address ranges and then collect a list of ports to check. You can initially check on port 443, but there are many ports on which certificates are commonly presented.
Many certificates are not discoverable via network ports, such as client-side certificates used for mutual authentication on SSL. Finding these certificates typically involves performing file system scans on server and client systems with a locally-installed agent.
Sounds simple! Just remember that performing an inventory is not a one-time event. You should repeat the steps above weekly to ensure the inventory is up to date.
As you’re developing your inventory, establish a correlation of who the contacts and owners are for certificates. Wherever possible assign groups as the contacts instead of individuals to avoid a single point of failure. Some helpful sources include certificate authorities, tracking spreadsheets, and even a CMDB. Define clear responsibilities for maintenance of certificate contact information.
An important method for preventing in-service expirations is to establish a central monitoring function that ensures certificates are replaced prior to expiration by automatically notifying responsible groups. Only when the new certificate has been installed and the application has been reset to use the new certificate prior to the time of expiration is the risk of downtime averted.
Establish standard practices for enrolment and provisioning that maximize reliability and repeatability, ensure security and compliance to policy, and minimize load on your administrators. There are easily 20 or more steps involved in issuing or renewing a certificate. These steps must be standardised and implemented in compliance with policy every time.
Everything you need to know about APIs for business
By Omar Javaid, president, Vonage API Platform, Vonage
If your work brings you into close proximity with technology, chances are that you’ve come across APIs. Like many of the tech acronyms we hear – DNS, VOIP, SaaS – APIs fall into a category of terms that most of us would consider best left to the IT department. However, APIs are a vital tool for any tech-enabled business, and a basic understanding of them at management level can help to drive sales, increase customer satisfaction, and improve the user experience.
Although they seem daunting, getting to grips with APIs is surprisingly straightforward. API stands for Application Programming Interface, and can be simply defined as a software tool used to control programmes. Essentially, APIs create sets of rules that allow applications to communicate with each other – they are the part of the server that receives requests and sends responses. Today, when data is transferred between a pair (or more) of programs or applications, an API normally makes it happen.
To give a real-world example: when a user types Instagram’s URL into their browser and hits the Return key, a request is subsequently transmitted to Instagram’s remote servers. That browser then processes the response code it receives and displays the page. For the browser, Instagram’s server is an API – allowing it to communicate and relay information back to you without interruption or delay.
The job of the API is to simplify the complex data exchanged between these servers, and to make the interaction as seamless as possible for the end user. Considering that the vast majority of our business and personal lives now take place virtually, any solution that optimises the online experience is extremely valuable.
Using APIs to improve the customer experience
One of the core benefits of APIs is that they enable businesses to free themselves from the time consuming and costly process of developing in-house software to power a single core application. Instead, developers can outsource certain tasks to remote “off-the-shelf” APIs, saving time, money, and allowing resources to be channeled elsewhere. These add-on services allow businesses to offer a more complete, one-stop solution to customers, whilst streamlining the process to optimise user experience.
Although we may not always realise it, APIs are playing a vital silent role in almost every purchase and interaction we have online. Take booking a holiday for example. As we browse comparison sights, APIs are working furiously behind the scene to aggregate information from airline databases, hotel websites, and excursion providers. The API performs the back and forth needed to retrieve the information, whilst we are able to sit back and view all of the results on the same page. Simplifying this process enables travel comparison websites to make the search for holidays quick and easy, and encourages customers to stay on the site by offering all that they need in one easy to consume package.
APIs also allow smaller businesses to utilise tools provided by some of the world’s largest and most successful companies. Google’s Calendar API for example could be used within a beauty salon website to enable customers to book and schedule treatment reminders, whilst Apple’s weather tool could be plugged-in to an events company website to give customers real-time weather updates. While the API’s developer does retain ultimate control over how the API is used, there are still countless ways to integrate these tools to benefit your business and improve the functionality of your website.
The recent Covid-19 pandemic in particular has highlighted the value of an API class that normally receives little attention; communication APIs.
Today, companies are boosting spending on unified communications-as-a-service (UCaaS), along with video conferencing, collaboration, and voice technology solutions given the exponential growth in home and remote working as a result. Where face-to-face contact is limited by necessity, businesses need to be able to communicate with employees and customers in ways which are secure, simple, and cost-effective.
Given how rapidly the technology landscape changes, APIs are the clear solution to avoiding the expense of developing tools from scratch, in addition to harnessing the power of the advanced features offered by established API providers.
Using them, businesses are able to adapt to suit changing customer preferences; for example offering an online chatbot to handle customer queries, or by using multi-channel messaging to connect with customers via WhatsApp or Messenger. These tools are not only useful, but can also allow you to gain intelligence into a customer’s preferences and habits – both useful marketing gauges.
On the other hand, comms APIs can also help to address problems that may crop up internally within organisations and workforces. There are APIs which allow callers to automatically sync calendars, meaning that meetings will only be scheduled when all parties can attend. There are also APIs for timezone conversion, permissions requests, and for video link calls and messaging. With the work from home trend continuing for the foreseeable future, investing in these areas is critical if businesses want to keep delivering at the highest levels.
Considering all of the above, it’s clear that we can expect to see the adoption of APIs continue. Developers are constantly working to create increasingly sophisticated products, and many have moved towards exclusively building and hosting APIs, rather than building the apps themselves – creating a so called “API Economy” of sorts.
This focus on creating the best possible APIs has allowed smaller businesses to harness the collective expertise of the world’s largest and most successful companies, and the chance to use these tools represents a fantastic opportunity for growth. The reach of APIs extends far beyond the IT department, and with a basic understanding, they can be used by senior management and leadership teams to optimise all areas of the business – not bad for three small letters.
Unexplained Wealth Orders: Rightly Celebrated or Over-Rated?
By Nicola Sharp of financial crime specialists Rahman Ravelli considers the attention given to unexplained wealth orders – and emphasises that they can be challenged.
There is little doubt that many sectors of the media – and their readers – enjoy a story that involves an unexplained wealth order (UWO). They do, after all, have many of the ingredients that many look for in a good tale: allegations of wrongdoing on a large scale, someone being made to hand over assets worth more than most people will earn in a lifetime and the sense that justice has been seen to be done.
In the latest UWO, which was widely covered in the media last week, Leeds businessman Mansoor Mahmood Hussain was compelled to hand over property worth just short of £10M, after being accused of acting as a money launderer. He has been ordered to surrender the assets because the National Crime Agency (NCA) believed his wealth was the proceeds of crime, and so considered him a suitable target for a UWO.
Introduced by the Criminal Finances Act 2017, UWOs give law enforcement agencies powers to require persons to explain how they came to possess their assets, and to show that their wealth has come from legitimate sources. A UWO can be sought without any civil or criminal proceedings having begun. There is no need for the subject of a UWO to have been convicted of an offence or to have had a civil law judgement against them. Agencies can apply to the High Court for a UWO against any property valued at over £50,000, if the person owning it is reasonably suspected of being involved in serious crime (or connected to a person who is) and there are reasonable grounds to suspect that a person’s lawfully-obtained income would be insufficient to allow that person to obtain that property.
Like Zamira Hajiyeva before him, Mansoor Hussain’s inability to provide a credible, innocent explanation for his wealth has cost him – and generated headlines. Hajiyeva may be best known for somehow racking up £16M of expenditure at Harrods. But this only became known when she was the first person to be the subject of UWOs. The NCA expected her to explain how she had bought a £11.5M Knightsbridge house and a £10.5M golf course in Ascot, bearing in mind her husband is the former head of the state-owned International Bank of Azerbaijan, had a salary of no more than $70,000 and was convicted of fraud and embezzlement. Earlier this year, she lost her appeal against the UWOs, thus enabling the media to re-run her story and giving the NCA the chance to make approving noises about UWOs being a valuable tool in tackling illicit finance.
But before there is a rush to applaud UWOs, it should be said that the NCA’s relationship with them has been a chequered one, to say the least. Since becoming available to the NCA, the agency’s success rate with UWOs has been patchy. This is despite the standard of proof for UWOs being significantly lower than that required in criminal cases. Last year saw the NCA granted three UWOs for London property valued at £80M. Yet less than a year later, these UWOs were discharged, with a judge criticising the NCA’s “unreliable’’ assumptions and “artificial and flawed’’ reasoning. The Court of Appeal then refused the agency permission to appeal this decision.
While a UWO is a tool that enables law enforcement agencies to seize assets they believe are the proceeds of crime without anyone ever being convicted, it does not yet appear to have become the great weapon against illicit wealth that many would have hoped. Of the four cases begun since UWOs were introduced, two are still being contested. Mansoor Hussain’s case is the first time a UWO has successfully led to the recovery of assets from an individual.
Although, a UWO can be seen as effective in certain situations, it will often be considered the most (and perhaps only) viable option when a prosecution has failed or when the authorities do not believe there is enough evidence for a realistic chance of a conviction.
When being faced with an UWO it should be remembered that whilst agreeing to settle and hand over property is not an admission of guilt, anyone facing a UWO must consider carefully how they respond to the authorities. It is vitally important to take the right advice. Deciding how to proceed when assets worth millions are at stake can be the biggest decision a person ever has to make.
In such circumstances it will often be the case that an intelligent, robustly-argued challenge to a UWO – and, in particular, to the allegations being made by the law enforcement agency seeking the UWO – will bring success. But that success will depend on knowing precisely how to respond – and who to turn to – if and when you become the intended target of a UWO.
How Siloed Data Leaves Financial Institutions Open to Fraud
By Stephany Lapierre, CEO Tealbook
Reducing the risk of fraud is a top priority for all financial institutions since fraud is responsible for massive profit loss, as well as the degradation of an institution’s integrity and brand.
In trying to prevent fraud, most executives look to protect themselves from the outside in, implementing layers of security and launching reactive measures. However, in order to truly protect your organization from fraud, it’s imperative to begin by looking at your existing internal structures. The most critical and often overlooked area to assess is how your organization obtains, enriches, and distributes data.
Streamlining and scrubbing your data can increase profitability without adding to resource spend. Having good data allows you to complete your due diligence on vendors and external entities your organization regularly deals with. It favorably adjusts your efficiency ratio and reduces risk by eliminating redundancies, conflicting information, and information gaps. In addition, it allows smaller teams to operate with increased scale and effectiveness. In turn, this leads to a more effective vendor vetting process and less room for error in payment information verification.
Conversely, poorly managed data is confusing and deceiving and can play an unfortunate role in giving fraudulent access to outside parties through internal miscommunications. For example, updates could be made in one system and not another, and suddenly different departments are working with different data sets like payment information or legal formation documents that regulators look for in audits, and no one knows what is true or accurate. This effect snowballs over time, creating massive holes in the integrity of the data, creating unnecessary risk exposure and audit failures.
All of these vulnerabilities can serve as the foundation for developing a risk management protocol that may be rendered useless if it is based on poor data. It is impossible to properly vet vendors and suppliers or verify payment information if the data is unreliable.
By investing in a solid Data Foundation, you’ll see an increase in the success of your risk management and fraud prevention measures. In many instances, you won’t need to add more steps or resources, just power your existing systems with clean, agile, and accurate data to see improved efficiency.
Here’s a closer look at the most common vulnerabilities within a typical financial institution’s data ecosystem:
Fragmented Organization Structure
As organizations grow and scale, it’s inevitable that different subsections will become isolated from one another and begin different processes for data management. Poorly managed systems can exacerbate this lack of communication and threaten data integrity.
It may not seem like cause for concern if a few different arms of an organization aren’t completely in sync. However, in the financial space, this issue rarely applies to just one or two organizational divides. For example, a prominent US-based financial institution boasts over 90 business units, all of which need to be synergized in order to prevent inaccurate data, redundancies, and problems with regulatory information gathering. This siloed information is, unfortunately, a common practice that needs to be addressed.
Unmanaged Proprietary Systems
In an attempt to serve data in a highly specialized way, many institutions have explored developing proprietary data systems for internal use. However, because of factors like employee turnover or an inability to keep up with data integrity best practices, these legacy systems quickly become obsolete and unmanaged. Their custom nature also renders them inflexible and unable to integrate with other solutions.
When trying to work around an unmanaged system, different branches of an institution may turn to different solutions. When work is being done across different platforms, this reduces visibility and increases risk for inaccuracies, which leads to poor decisions, costly rework, and potentially fraud.
If your organization is reliant on a proprietary system, consider if that system is functional and scalable. If it’s not, you may want to look into a flexible data management system that can work with other technologies.
Disparate Information Across Systems
Mergers, acquisitions, and growth also lead to using and implementing many different ERP solutions and antiquated legacy software that are forced to communicate with each other using painful manual efforts. A major problem arises from the fact that these systems operate across numerous lines of businesses, all with different siloed data. By having so many siloed systems that could be compromised with harmful data, these disparate data sources leave banks and other financial institutions exposed to unnecessary risk.
Different departments have different needs, so it makes sense that they would use different solutions, but it’s important that those solutions pull from a single source of truth in order to prevent the types of data inaccuracies that lead to vulnerabilities.
Closing the holes in your data integrity is the most proactive way a financial institution can defend against fraud. As hackers get increasingly creative and aggressive, it becomes even more critical that organizations have a trusted Data Foundation to base their decisions on. This can be achieved by ensuring that siloed systems are powered by consistent and accurate data from a single reliable source.
Adoption of tech in private markets lags behind industry trends
Nine out of ten financial institutions have accelerated their digitisation strategy as a result of Covid-19. Yet just 26% of...
Covid-19 disruption drives five new retail supply chain trends
The business disruption caused by COVID-19 has resulted in four out of five (82%) retailers changing their approach to stock...
Remote leadership anxieties
It’s a difficult time to be navigating the complex world of business. Whilst adapting to new ways of working remotely,...
Online jobs soar by 14% in third quarter 2020, Freelancer.com’s Fast 50 reports
Freelancer.com (ASX: FLN), the world’s largest freelancing and crowdsourcing marketplace by number of users and jobs posted, today released the...
One third of money management tools face closure by the end of the year if they do not embrace open banking
New research from Yolt Technology Services shows 35% of Personal Finance Managers aren’t using any open banking technology Imminent screen...
Pivoting growth strategy to rebuild consumer trust and confidence
By Richard Steggall, the CEO of Urban FT Trust is essential to all relationships, whether personal or professional. And in...
Everything you need to know about APIs for business
By Omar Javaid, president, Vonage API Platform, Vonage If your work brings you into close proximity with technology, chances are...
Accountants have become critical to the survival of businesses and their reputations during Covid-19
The opportunity for fraudulent activity to flourish as finance departments operate remotely with less oversight in these extraordinary Covid-19 times...
Unexplained Wealth Orders: Rightly Celebrated or Over-Rated?
By Nicola Sharp of financial crime specialists Rahman Ravelli considers the attention given to unexplained wealth orders – and emphasises...
Taking advantage of the UK’s renovation revolution
By Paresh Raja, CEO, Market Financial Solutions UK property is a popular asset class because of its historical resilience to...