Updating an existing database (either CRM data or a data product) is a never-ending job that most people would like to avoid. We built Information Evolution primarily to address publishers’ need to keep on top of the shifting technology trends that can make this expensive and often thankless task as inexpensive, quick, and simple as possible.
Database owners can take two fundamentally different approaches to updates: direct primary source updating and data overlays. The former is more expensive and guarantees timeliness and accuracy. The latter
is cheap and fast. We prefer primary source research to quickie overlays and here’s why.
Overlaying third-party data on your database has several thorny issues:
- Is the database to which you are comparing your data as accurate as possible? This is unlikely for most B2B publishers who are usually the best sources of deep information on given industries. Also it implies that the overlaid data has an almost “God-like” level of accuracy. In other words, the data is an accurate reflection of “reality.”
- What, really, is the provenance of the data you are overlaying? Large “God-like” databases are often the result of massive aggregations of information, each “chunk” of which has a different age and source. Some sources are murky at best and the result of a dubious harvest at worst.
- What do you do with the results of an overlay? If you unilaterally accept every case where the overlaid data is substantially different (and there are a lot of way to define that) then the process is indeed quick. But did you just overwrite a newer record with old data? There’s no way to know without accurate provenance information or primary research. And what about the non-matches? Do you just delete them?
In the end, IMHO, neither option is ideal. Monitoring and maintaining the data in real time needs to be the foundation of any data maintenance effort. This involves data harvesting and change detection via tools like Connotate and Enlyton.
Telephone verification is also still an essential part of the process. Adding new fields of data to company records during every time a primary source research effort can offset this expense. Bringing in another couple of contact names and emails, appending deeper info on the vendors that companies use, or adding social media presences for firms are all effective ways data owners can make updating both effective and profitable.
posted by Shyamali Ghosh on December 18, 2014
by Matt Manning
The recent DataContent track at the BIMS conference in Miami last week pointed to some very clear trends in data product development. Here are some of the memes we’re likely to hear a lot more about in 2015.
Predictive Analytics: The ability to predict who is in the market to buy different types of products and services is truly coming of age.
- Cooperative models like MadisonLogic let members overlay their customer data with aggregate data from other co-op members. This immediately identifies high-value prospects (see “lead scoring” below).
- Penton can now identify new products and product categories just as they’re trending. They then sell this data as part of high-end analytical products priced from $5-$15,000/year.
- Walmart uses data on regional events, weather, and other factors to accurately forecast product demand.
- Russell Perkins reminded attendees that having an accurate underlying universe definition is key to the success of any predictive product.
The API Economy: Salesforce and other CRM integration drove much of the initial market interest in APIs. The real trend is now toward building API-based data products rather than full-blown online information services (also known as “Data as a Service”).
- Edmunds, KnowWho, OpenCorporates and many others are firmly on the API bandwagon.
- Those offering free access to low-volume use of their APIs are seeing extraordinary success and spurring more applications using their data.
- When it’s done right, APIs allow easy workflow integration (which is the holy grail). They also reduce the need for expensive and complex interfaces, security protections, and transactional systems.
Social Integration: Events are becoming the crown jewels in publishing company properties. It’s partly due to their social dimensions and the way that so much information and brand marketing revolves around events.
- Equilar is aggressively pursuing the “event networking” market for its incredibly detailed info on C-level execs.
- Social3 offers the ability to see which customers have the most social clout with their new social lead-scoring service.
- Reachable’s product offering mines the social networks of employees and business partners to uncover personal connections to high-value prospects.
Lead scoring: It is becoming common to not just sell access to an audience (via a list sales, ad sales, etc.), but to rank prospects so customers’ sales efforts are faster and more successful.
- The best leads are prospects who buy a lot, buy from multiple competitors, and buy the same way that you market to them.
- Editorial departments are using popularity ranking to drive coverage in much the same way as content marketers do.
Discovery advertising: The art and science of displaying teasers for related articles to “engage” audiences and earn ancillary revenues has matured into a serious business segment. Taboola will earn $250 million in 2014 from these services.
“Renting” data: Craig Harris from HG Data made the case for an alternative to the subscription model: All information delivered to a customer must be removed from their systems if they do not renew their subscription to the data. This is another clever way to permanently embed data services in the workflow.
Deeper Data: Entrepreneurs continue to find new “deep” data sets to add value to company, people, and product databases.
- LowerFees.com bases their new hyperlocal information service on price data for consumer service companies like oil change centers, dentists, etc.
- Nutritionix is building a massive database on the nutritional value of all food products sold in the US.
Driving Usage: This remains a hot topic especially as services get more complex and costly.
- D&B’s Data Exchange is a great cooperative approach to driving sales for industry-specific B2B publishing partners. It also forces them to append the proprietary DUNS ID numbers, thus driving D&B sales.
- ScoutAnalytics: helps publishers monitor subscription service use and proactively drive more usage, improving renewal rates.
posted by Shyamali Ghosh on November 24, 2014