Using a search engine for financial risk assessment won’t give you the results you need
Comprehensive due diligence is more important now than ever. The new normal brought about by COVID-19 has significantly impacted individuals and businesses. As a result, financial institutions have dealt with increased business due to the CARES Act and the issuance of Payroll Protection Program loans.
However, this uptick in business for financial institutions has also meant an increase in opportunity for fraudulent activity. With so much data and information seemingly at our fingertips, it is easy to think that vetting a new customer or checking the reputation of a business can be done through your favorite search engine.
This simplicity is a deception.
The truth is, search engines are not reliable tools for meaningful due diligence, particularly when it comes to identifying suspicious financial activity, fraud, money laundering, criminal involvement, or other forms of legal malfeasance.
In fact, organizations relying on mainstream search engines for fraud protection and regulatory compliance may be unwittingly exposing themselves to dangerously high levels of risk. In addition to its technical limitations, one of the worst things a basic internet search can do to a financial institution’s stakeholders is fool them into believing they have done their due diligence when they haven’t.
Not all search algorithms are the same
To understand why conventional search engines are such an inadequate tool for uncovering critical financial risk factors—and why targeted software solutions developed specifically for fraud prevention are so much more effective—it helps to understand some of the capabilities of these search engines.
A search engine’s bots—called “spiders” or “crawlers”—follow links on the internet and index what they find. Every time a user conducts a query, that search engine employs a series of algorithms to pull relevant content from within its index and display it—ranked according to what it thinks is most relevant for that specific search.
In order to identify the web pages most likely to satisfy a search query, search engine algorithms use several indicators, including term relevance, credibility of the website, usability, location, etc. Of course, another factor is that others conducting similar searches have found those pages useful, too.
However, search engines can only provide results for pages to which they have full access. Websites or sections within them that require logins—or some kind of information to be entered in order to access them—aren’t going to show in the search engine results pages (SERPs). They also can’t return sites or pages that aren’t indexable.
More isn’t always better
More is better in many ways, but the great paradox of big data is more information does not necessarily mean better information—in fact, it’s quite the opposite. The ease with which certain types of information can be found is often inversely proportional to the amount of information available. When the Google search engine was introduced in 1998, there were about 2.4 million websites on the internet. Now there are more than 1.5 billion websites worldwide containing 33 zettabytes of data or 33 trillion gigabytes. By 2025, the International Data Corp. (IDC) estimates that number will explode to 175 zettabytes and grow exponentially from there.