Wednesday, October 3, 2007

IBM 'continuum' helps companies crawl-walk-run along the SOA path

IBM today unleashed a barrage of announcements that cover services and software to enable a crawl-walk-run strategy for enterprises, as they move into the services-oriented architecture (SOA) world.


The new offerings range from a "SOA Sandbox," which will allow developers to "play" with SOA before using it in a production environment, to WebSphere enhancements, and finally, to an unveiling of the new IBM Optim, a data governance application from recently acquired Princeton Softech.

Key to the crawl-walk-run approach is what IBM calls their SOA Continuum, which takes companies from the very basic -- involving only focused, high-ROI projects -- to the most advanced, in which technology becomes invisible and more than 80 percent of business functions are delivered as services -- and more than 50 percent of those are re-used.

I sort of remember the logic from an earlier blog. Glad they agree.

Recognizing the need for what they've dubbed the Globally Integrated Enterprise, IBM has also introduced assessment tools, including a Benchmark Wizard, which is based on key agility indicators and is loaded with 1,100 qualitative business indicators and uses best practices derived from 1,600 case studies. This will allow an enterprise to determine how it stacks up against the industry in general.

The new offerings also address several key stumbling blocks along the path to SOA-based agility. One such sticking point is how systems respond when services from various sources may be unavailable. IBM says its updates WebSphere Process Server has extensive compensation support, allowing process to recover reliably when target applications are unavailable.

Another key concern is exposing sensitive or personal data when services access databases. The newly acquired Optim product is designed to identify and protect private data in complex application environments.

IBM also announced enhancement for process integrity to several products, including:

  • Message Broker and MQ
  • Tivoli Composite Application Manager for SOA
  • WebSphere DataPower XML Security Gateway
  • IBM Information Server

IBM is taking an "all things to all people" approach and offering new SOA configurations, designed to help enterprises use legacy and packaged applications in a SOA environment. These include best practices and step-by-step implementation guides.

To help developers get started -- before even starting to crawl along the SOA path -- IBM has set up a SOA Sandbox, a free test bed on the developerWorks site. The sandbox, where potential users can download trial software or can "play" online in a hosted environment, includes software, tutorials, quick-start guides, and best-practices guidance.

As another aid to companies, IBM is offering "SOA Healthchecks," workshops to assess and diagnose applications, services and infrastructure to determine application reuse, security, and infrastructure flexibility.

Look for this opening salvo to increase in crescendo over the next few months, as IBM rolls out these and more SOA offerings at 500 events, including a customer Webcast on Oct. 9, and a series of announcements to create a drumbeat about the latest SOA end-to-end initiative.

Engagement on SOA has to take place in countless ways. IBM seems to be up to the task, or least is ready to try.

Monday, October 1, 2007

Adobe pushes foward on enabling RIA ecosystem with Flex 3, Adobe AIR betas

The world of rich Internet applications (RIA) got a little richer on Monday when Adobe Systems released the beta versions of Flex 3 and the Flex SDK with enhanced features for evelopers.

The beta versions are in place of the alpha release, which was scheduled for this month, but which has been pushed out to the first quarter of 2008 to coincide with the release of Adobe AIR 1.0. The AIR beta is also available.

Enhanced capabilities of the Flex beta include support for ASP.NET, including new data wizards that allow developers to look at data tables and create a new Flex application from a SQL database. Developers who consume Web services can view the WSDL files and automatically generate code for invoking operations.

Adobe AIR now supports background applications and system tray notifications, allowing the app to run in the background. Among its other features are:

  • Synchronous APIs for embedded local database
  • Greater control of windows and menus
  • Content protection for video, HTML improvements
  • Application and runtime enhancements such as improved install process and automatic updates for the runtime.

Check out Matt Chotin's blog at Adobe for a complete rundown on the latest enhancements and updates.

Also, keep an eye out for Adobe to announce a new Adobe Developer Connection, where the Adobe community can get newsletters, form a professional network, participate in forums, and get special offers.

The Adobe AIR and the Flex betas can be downloaded from the Adobe Labs site, which also has a pre-release version of Adobe Media Player, which will be free to end users and will allow companies to distribute Web 2.0 content in video format. Major television broadcasters and other content producers have already agreed to support the Adobe format.

Adobe made the announcements at Adobe Max 2007. Among other news from the event:

Adobe has signed a definitive agreement to acquire Virtual Ubiquity and its online word processor, Buzzword, built with Adobe Flex and leveraging Flash Player. It's soon to be available on Adobe AIR. Buzzword has integral collaboration capabilities that allows multiple authors to edit and comment on documents from anywhere at anytime. Because it will run on Adobe AIR, the application will offer users a hybrid online/offline experience for working with documents.

Adobe and Business Objects have agreed to jointly undertake multiple initiatives to drive product interoperability and optimization, technology adoption, and product distribution.

One of the key initiatives is the development of a Business Objects Xcelsius Connector to Adobe LiveCycle Data Services ES.

Adobe is posing a strong alternative to other platform approaches to RIA and Enterprise 2.0 interfaces, most notably against Microsoft. It's just one of the thousands of daggers pointed at Redmond nowadays, but its cuts could go deep.

I'd like to see Adobe make more partnerships with those vendors supporting SOA activities. More ease and integration of RIAs and SOA infrastructure could be powerful. Microsoft obviously thinks so. I wonder why IBM and Adobe are not closer, at least in their more blatant go-to-marker campaigns.

I'd also like to see Adobe move more quickly on the completeness of the platform approach. These RIAs are catching on fast.

Software AG consolidates governance products for SOA

Software AG took the "big picture" view of governance for service-oriented architecture (SOA) with the announcement today of the CentraSite Governance Edition.

The approach aims to enable governance more broadly across the enterprise, with policy enforcement capabilities building on the webMethods Infravio X-Registry (acquired by SAG with Infravio earlier this year). Enforcement features are built directly into the platform, which offers a richer and more extensible metadata repository, a more intuitive Web2.0 interface, and wizard-driven templates.

While SOAs can help companies achieve much-needed business agility -- by flexibly arranging and repurposing existing and new IT assets, both from within and outside the enterprise -- that agility comes at the price of greater complexity.

One stumbling block has been the governance of SOA assets and processes, along with the ability to set, adjust and enforce policies. SOAs will require finer management of assets as they access and expose ever more enterprise data and interactions.

Encompassing both a UDDI (Universal Description, Discovery and Integration protocol), v. 3.0 compliant registry and a fully extensible, JAXR complaint (Java API for XML Registries) repository for maintaining associated metadata and policies, CentraSite Governance Edition can be implemented as the design-time, run-time and change-time governance platform for any heterogeneous environment.

It also acts as a policy hub for run-time enforcement with an integrated run-time policy enforcement point, webMethods X-Broker, and standards-based support for additional third-party applications used to mediate transactions betweens service providers and consumers.

A new feature in CentraSite Governance Edition, available now, is Active Policy, which automates SOA processes and simplifies end-user adoption. It is also pre-loaded with more than 80 pre-defined best practices that help streamline end-user adoption.

Other features include:

  • Unified lifecycle governance
  • Enhanced repository
  • Open metadata model
  • Enhanced change-time governance
  • Customized views
  • SOA federation
  • Standards support

Tony Baer at CBR Online sees Software AGs move along its webMethods roadmap as being "right paced" and even takes a look into the future:

"Software AG is moving very deliberately to execute on the roadmap for converging products from the acquired webMethods. With announcements right paced at about every four weeks, this is the second such announcement to come out of Software AG since it unveiled the roadmap back in August.

"The company is also being upfront, in that, while it is making the products look cosmetically the same at the UI level and in branding, it admits that the real convergence will happen next year. At this point, both have, or will have, common Ajax-based rich user interfaces. Both rely also on the JAR (Java API for XML Registries), so access in also via a common method."

Tuesday, September 25, 2007

Integration infrastructure approaches adjust to new world of SaaS and shared services

Read a full transcript of the discussion. Listen to the podcast. Sponsor: Cape Clear Software.

Change is afoot for the role and requirements of integration for modern software-as-a-service (SaaS) providers and enterprises adopting shared services models. Reuse is becoming an important issue, as are patterns of automation.

The notion of reuse of integration -- with added emphasis on integration as a service -- has prompted a different approach to integration infrastructure. The new demand is driven by ecologies of services, some from the Web "cloud," as well as the need to efficiently scale the delivery of services and applications composed of many disparate component services.

Integrations require reusable patterns, high performance, as well as many different means of access from clients. As a result Cape Clear Software has this week unveiled a new major version of its enterprise service bus (ESB), Cape Clear 7.5, with an emphasis on:
  • A new graphical editor, the SOA Assembly Editor, an Eclipse-based tool to graphically clip together elements of integrations.
  • Multi-tenanting additions to the ESB that allow segmentation of integrations, data, and reporting, as well as segmenting use and reuse of integrations on reporting and management of integrations based on the identities of inbound customers, clients, or businesses.
  • A Business Process Execution Language (BPEL) management system with tools to monitor transactions, and repair transactions when they fail, to allow for rebuilding previous business information and ensure transactional integrity in running and maintaining large enterprise-class BPEL deployments.
To help better understand the new landscape for integration models, I recently moderated a sponsored podcast discussion with Phil Wainewright, an independent consultant, director of Procullux Ventures, and fellow ZDNet SaaS blogger, as well as Annrai O’Toole, CEO of Cape Clear Software.

Here are some excerpts:
... We're getting more sophisticated about SaaS, because it's being taken on board in a whole range of areas within the enterprise, and people want to do integration.

There are two forms of integration coming to the fore. The first is where data needs to be exchanged with legacy applications within the enterprise. The second form of integration that we see -- not at the moment, but it’s increasingly going to be an issue -- is where people want to integrate between different services coming in from the cloud. It’s a topic that’s familiar when we talk about mashups, fairly simple integrations of services that are done at the browser level. In the enterprise space, people tend to talk about composite applications, and it seems to be more difficult when you are dealing with a range of data sources that have to be combined.

People have realized that if you're doing integration to each separate service that's out there, then you're creating the same point-to-point spaghetti that people were trying to get away from by moving to this new IT paradigm. People are starting to think that there's a better way of doing this. If there's a better way of delivering the software, then there ought to be a better way of integrating it together as well.

Therefore, they realize that if we share the integration, rather than building it from scratch each time, we can bring into the integration field some of the benefits that we see with the shared-services architecture or SaaS. ... The new generation of SaaS providers, are really talking about a shared infrastructure, where the application is configured and tailored to the needs of individual customers. In a way, they’re segmented off from the way the infrastructure works underneath.

When you build an integration, you always end up having to customize it in some way for different customers. Customers will have different data formats. They’ll want to access it slightly differently. Some people will want to talk to it over SOAP. Some won't, and they’ll want to use something like REST. Or they might be going backwards and are only able to send it FTP drops, or something like that.

Multi-tenanting is one solution to the problem. The other is what we call multi-channel, which is the ability to have an integration, and make it available with different security policies, different transports, and different transformations going in and out.

A combination of multi-tenanting and multi-channeling allows you to build integrations once, make them accessible to different users, and make them accessible in different ways for each of those different customers. It gives you the scalability and reuse you need to make this model viable.

One point worth bearing in mind here is that this problem is going to get solved, because the economic reality of it suggests that we must solve this. One, the payoff for getting it right is huge. Second, the whole model of SaaS won’t be successful, unless we skin the integration problem. We don’t want the world to be limited to just having Salesforce.com with its siloed application.

We want SaaS to be the generic solution for everybody. That’s the way the industry is going, and that can only happen by solving this problem. So, we’re having a good stab at it, and I'll just briefly address some of the things that I think enable us to do it now, as opposed to in the past. First, there is a standardization that’s taken place. A set of standards has been created around SOA, giving us the interoperability platform that makes it possible in a way that was never possible before. Second is an acceptance of this shared-services, hosted model.

Years ago, people would have laughed at you and said, "I’m going to trust all my customer data to a provider in the cloud?" But, they’re doing it happily because of the economics of it. The whole trend toward trusting people with outsourced offerings means that the people will be more likely to trust integrations out there, because a lot of the technology to do this has been around for quite some time.

In enterprises you’re seeing this big move to virtualization and shared services. They’re saying, "Why are we having development teams build integration in all these branch offices at all these locations around the world? It’s extremely wasteful. It's a lot of skill that we've got to push out, and there are a lot of things that go wrong with these. Can't we consolidate all of those into a centralized data center? We’ll host those integrations for those individual business units or those at departments, but we'll do it here. We’ve got all the expertise in one place."

Those guys are delighted, because at the individual local level they don’t maintain all the costs and all the complexity of dealing with all the issues. It’s hosted out in their internal cloud. We haven't seen enough data points on that, but this hosted integration model can work. We’ve got it working for pure entities in SaaS companies like Workday, and we’ve got it working for a number of large enterprises. There is enough evidence for us to believe that this is really going to be the way forward for everybody in the industry.
Read a full transcript of the discussion. Listen to the podcast. Sponsor: Cape Clear Software.

Monday, September 24, 2007

Integrien deepens analytics, betters interoperability and userability in Q4's Alive 6.0 release

Read a full transcript of the discussion. Listen to the podcast. Sponsor: Integrien Corp.

The movement of IT and systems management to the end-to-end business service value level has been a long time in coming. Yet the need has never been higher. Enterprises and on-demand application providers alike need to predict how systems will behave under a variety of conditions.

Rather than losing control to ever-increasing complexity -- and gaining less and less insight into the root causes of problematic applications and services -- operators must gain the ability to predict and prevent threats to the performance of their applications and services. Firefighting against applications performance degradation in a dynamic service-oriented architecture (SOA) just won't cut it.

By adding real-time analytics to their systems management practices, IT operators can determine the normal state of how systems should be performing. Then, by measuring the characteristics of systems under many conditions over time, administrators can gain predictive insights into their entire operations, based on a business services-level of performance and demand. They can stay ahead of complexity, and therefore contain the costs of ongoing high-performance applications delivery.

I recently had a podcast discussion with Mazda Marvasti, the CTO of Integrien Corp., on managing complexity by leveraging probabilistic systems management and remediation. I learned that Integrien's Alive suite uses probabilistic analysis to predict IT systems problems before costly applications outages. Furthermore, I received some details on the next Alive 6.0 release in Q4 of this year.

Here are some excerpts:
Can you give us some sense of the direction that the major new offerings within the Alive product set will take?

Basically, we have three pillars that the product is based on. First is usability. That's a particular pet peeve of mine. I didn't find any of the applications out there very usable. We have spent a lot of time working with customers and working with different operations groups. ... The second piece is interoperability. The majority of the organizations that we go to already have a whole bunch of systems, whether it be data collection systems, event management systems, or configuration management databases, etc.

Our product absolutely needs to leverage those investments -- and they are leveragable. But even those investments in their silos don’t produce as much benefit to the customer as a product like ours going in there and utilizing all of that data that they have in there, and bringing out the information that’s locked within it.

The third piece is analytics. What we have in the product coming out is scalability to 100,000 servers. We've kind of gone wild on the scalability side, because we are designing for the future. Nobody that I know of right now has that kind of a scale, except maybe Google, but theirs' is basically the same thing replicated thousands of times over, which is different than the enterprises we deal with, like banks or health-care organizations.

A single four-processor Xeon box, with Alive installed on it, can run real-time analytics for up to 100,000 devices. That’s the level of scale we're talking about. In terms of analytics, we've got three new pieces coming out, and basically every event we send out is a predictive event. It’s going to tell you this event occurred, and then this other set of events have a certain probability within a certain timeframe to occur.

Not only that, but then we can match it to what we call our "finger printing." Our finger printing is a pattern-matching technology that allows us to look at patterns of events and formulate a particular problem. It indicates particular problems and those become the predictive alerts to other problems.

Now, with SOA and virtualization moving into application-development and data-center automation, there is a tremendous amount of complexity in the operations arena. You can’t have the people who used to have the "tribal knowledge" in their head determining where the problems are coming from or what the issues are.

The problems and the complexity have gone beyond the capability of people just sitting there in front of screens of data, trying to make sense out of it. So, as we gained efficiency from application development, we need consistency of performance and availability, but all of this added to the complexity of managing the data center.

That’s how the evolution of the data center went from being totally deterministic, meaning that you knew every variable, could measure it, and had very specific rules telling you if certain things happened, and what they were and what they meant -- all the way to a non-deterministic era, which we are in right now.

Now, you can't possibly know all the variables, and the rules that you come up with today may be invalid tomorrow, all just because of change that has gone on in your environment. So, you cannot use the same techniques that you used 10 or 15 years ago to manage your operations today. Yet that’s what the current tools are doing. They are just more of the same, and that’s not meeting the requirements of the operations center anymore.

I’ve been working on these types of problems for the past 18 years. Since graduate school, I’ve been analyzing data extraction of information from disparate data. I went to work for Ford and General Motors -- really large environments. Back then, it was client-servers and how those environments were being managed. I could see the impending complexity, because I saw the level of pressure that there was on application developers to develop more reusable code and to develop faster with higher quality.

The run book is missing that information. The run book only has the information on how to clean it up after an accident happens.

That’s the missing piece in the operations arena. Part of the challenge for our company is getting the operations folks to start thinking in a different fashion. You can do it a little at a time. It doesn’t have to be a complete shift in one fell swoop, but it does require that change in mentality. Now that I am actually forewarned about something, how do I prevent it, as opposed to cleaning up after it happens.
Read a full transcript of the discussion. Listen to the podcast. Sponsor: Integrien Corp.

Before we know who owns the SOA business case, how about simple business processes?

There's a good article on "owning" the business case for SOA on SearchWebServices.com. Some of my most respected analysts are quoted.

But is the question posed a relevant one? While making the business case for SOA is and will be a fascinating topic for some time, we may be jumping the gun.

From where I sit, just about everyone that has a strategic role in IT and business decisions at an enterprise has an "ownership" stake in SOA. It's that pervasive. The COO may be the best person under many current organizational charts to see all the moving SOA parts.

Yet buy-in and inclusiveness -- both wide and deep -- for SOA are essential, so it can't really fall to any one person. Assigning "business value" ownership is too abstract, really, for real-world companies to begin using it and embracing SOA. SOA is ubiquitous in its effects. The positioning of SOA as an abstraction is holding back its embrace and adoption.

So let's look at more practical questions on SOA and business value, before we go shooting for the moon. Sadly, in even the most progressive enterprises, the ownership of a single business process is ambiguous. Organizations have been ceated for decades based on the notion of decentralization -- which is just another way of breaking up complexity into small chunks and assigning responsibility for the chunks, often at the expense of minding the whole. Very few individuals or teams are defined or incentivized to manage an entire business process. Yet this an essential stepping stone to SOA, and to eventually making the business case for SOA.

We see attempts to proffer SOA from the top down, with even less emphasis on adoption from the bottom up. What I'm saying is also, and perhaps predominantly, build it from the middle out. Create the new middle for SOA at the business process level, and then evangelize it in any which way.

In effect, SOA and its foundational core, business processes, are fighting back against the long-term tide of decentralization and IT specialization. SOA says you can now make the chucks of discrete IT resources relate far better, so why not begin to look at an entire process and work to make it more efficient, and more flexible? Why not extract the best of specialization and improve, refine and reuse the parts best in the context of general business -requirements whole? See the forest and the trees. Make better business decisions -- operationally and strategically -- as a result.

As Dr. Paul Brown points out in a book I recently helped review via a sponsored podcast, Succeeding with SOA: Realizing Business Value Through Total Architecture (Addison-Wesley, April 2007), the business process is the right level to assign "ownership." Now.

When an analyst or architect -- as well as their teams -- begin to see themselves as managing and evangelizing on a business process level, then SOA can begin to make strides as a concept and methodology more broadly. To try and inject SOA into a company broadly, then discretely is putting the cart in front of the horse. Better yet to re-arrange all the horses and carts based on the right trips for the right loads, making it easier to change horses and carts as needed.

I like the idea of cross-functional teams (horses, carts, drivers, and caravans) created that serve a business process lifecycle. These would be pods (perhaps virtual in nature) of tightly-coordinated people with the right mix of skills and experience -- specific and general, technical and business-oriented, able to communicate as a team on many levels.

Like the Ray Bradbury book, Fahrenheit 451, where individuals learn and carry on whole books in their memories as a way to preserve the books and their knowledge, business process pods would retain and refine the essence of a business process and care for it and extol its virtues throughout an enterprise. They would cross all the chasms across the constituent services but at the higher business value level.

We've heard talk of a "T" person from SOA evangelists at IBM, whereby the horizontal bar in the "T" represents business acumen, and the vertical bar represents technical depth. But I like the idea of the cross-functional pod better -- a team of, by, and for the business process.

The ownership of a business process (never mind SOA) is too much for one person. A multi-talented team can provide the wetware and organizational dynamism to get SOA started on a practical, middle level -- that of a business process as a productivity entity. This step is what's needed before we start assigning ownership for the business case for SOA.

Friday, September 21, 2007

Developing enterprise applications for mobile devices remains way too hard

A logjam exists between developers and their ability to productively deliver enterprise applications and data to mobile devices, such as cell phones, PDAs, and so-called converged devices like the Apple iPhone.

The logjam is complexity and too many obnoxious variables. To develop applications that reach even a small number of major handset environments means big-time custom plumbing, from the various data sources, to the mixture of networks, to the choices on synchronization, to the various security needs, to the many user interfaces and mobile client operating systems. Managing all these variables requires a high degree of skill across many different skill sets. There are not many developers that fit this bill in your average enterprise.

And this all means a lot of time and money is required to bring just a few basic applications to just a few basic mobile clients. No wonder enterprise mobility stubbornly remains below the radar for IT leadership. Mobile remains relegated to the crowded back burner of IT imperatives.

And given all the variables and high degree of required customization, few ISVs have emerged to try and make a living at producing mass-market mobile applications. The subsequent lack of killer applications, other than standards-based email and text messaging, reduces the appetite to take on the infrastructure complexity for taking the corporate datacenter out to the mobile client across commercial mobile networks.

Also impacting the complexity is the diversity in how mobile devices work from geographic and regulatory market to market. It's nearly impossible to envision a global approach to mobile wireless computing, as we've seen for desktop and web computing. Designing for one mobile market does not give you much of a leg-up in reaching many others.

The "inclusive platform" approach of dictating the exact device and/or runtime environment up and down the stack that mobile applications play in is itself stifling given the inability to take advantage of the low-cost devices and services available via commercial mobile service carriers. What's more, aligning the back-end and front-end infrastructure does not necessarily align with how the mobile telecommunications carriers and handset operators, well ... operate. They don't like the idea of losing control of what their clients do on their networks.

Clearly, wireless handheld delivery of enterprise data and applications has yet to reach its untapped and vast potential.

Microsoft for years has been grappling with these issues, with many fits and starts. There have been some impressive successes with Windows Mobile, but Microsoft has by no means sewn up the field of mobile enterprise applications design and delivery. Microsoft with its Windows Mobile approach has not yet achieved a critical mass for how business applications and data can be driven out to a field-based workforce. And it's likely that today's widely heterogeneous environment for end points and devices will be with us -- at least in the U.S. -- for a long time to come.

These seemingly intractable roadblocks to wider mobile business use are one reason we're seeing what amounts to appliances for the client devices. The Apple iPhone, which launched with great fanfare in late June, is a prime example. Apple has fused hardware, software and applications -- as well as a few critical APIs -- and has picked one opening-inning carrier, AT&T, as well as Wi-Fi generally, for connectivity. And that combination of attributes, along with what should soon be more APIs, makes the converged device/appliance approach not just a consumer affair -- it begins to help reduce the complexity for enterprise mobility too.

Microsoft in the summer said that it likes the idea of such fused mobile clients -- the appliance on the client -- so much that it has hinted it will produce such integrated devices too. Rumors persists that Google also has its sights on a mobile handheld appliance of some sort. Things are clearly heating up.

These nifty clients should go quite a way to making enterprise mobility far easier by allowing developers to exploit them with fewer interface and connectivity variables to manage. But the converged client still needs a back-end or middleware counterpart to help coordinate an enterprise's data, logic, and security needs. Sybase is hard at work on what may prove to be a game changer for such enterprise mobility middleware -- especially when coupled with a converged device such as the iPhone.

Sybase has bet its future growth on mobility. And while the Dublin, Calif. company has not yet announced the details of its full stack, the Sybase vision makes a lot of sense. Expect in the next year to see what amounts to a distributed middleware system approach to enterprise mobility from Sybase that may break the current logjam in enterprise mobile development and deployment. If Sybase can shake up the enterprise mobility infrastructure industry -- and partner effectively with the likes of Apple's iPhone -- then Microsoft's response will need to be swift and significant.

Dare I say it? What the heck: Look to 2008 to be the year that enterprise mobility finally gets some legs.

Monday, September 17, 2007

Survey uncovers heightening reliance on search across business purchasing

Listen to the podcast. Or read a full transcript. Sponsor: ZoomInfo.

It seems that businesses, whether they're small or global 2000 concerns, are buying more supplies using search at some point in the B2B procurement process. Some people begin and end a procurement journey with search. They actually buy the products through a strictly search-dependent process.



Yet many still use a combination of word-of-mouth, search, and traditional information gathering to guide them to the best deals on the most goods.

To find out just how much B2B buying behaviors are shifting, Enquiro Search Solutions conducted a survey earlier in 2007. They found that online search was consistently employed throughout the entire buying process, from awareness right through to purchase.

There’s still a lot of back and forth: Offline factors influence online activity, and vice-versa, for a merging of the online and the offline worlds. In an audio podcast discussion, as well as the accompanying BriefingsDirect multi-media video-podcast, I helped plumb the depths of Enquiro's findings and then vetted them through the experiences of B2B search engine ZoomInfo.

Join Gord Hotchkiss, President and CEO of Enquiro, and Bryan Burdick, COO of ZoomInfo, with moderation by myself, Dana Gardner, for a deep dive on B2B search trends and analysis.

Here are some excerpts:
We did the original survey in 2004 and, at the time, there wasn't a lot of research out there about search in general, even on the consumer side. There was virtually nothing on the B2B side. The first survey ... certainly proved that search was important. We found that online activity, in particular that connected with search activity, was consistent in a large percentage of purchases. In 2007, we added more insight to the methodology. We wanted to understand the different roles that are typical in B2B purchases -- economic buyers versus technical buyers versus user buyers. We also wanted to get more understanding of the different phases of the buying cycle.

As far as the main takeaways from the study, obviously online activity is more important than ever. In fact, we asked respondents to indicate from a list of over 30 influencers what was most important to them in making the purchase decision. Online factors, such as interaction with the vendor Website and interaction with the search engine were right up there with the traditional winner, word of mouth. What we see is a real link between those and looking for objective information and specific detail.

We did notice an evolution of behavior as you move through the funnel, and the nature of the interactions with the different online resources changes how you navigate to them and how you go to different sites for information. But, online research was consistent through the entire process, from awareness right through to purchase. There’s a lot of back and forth. ... We saw a merging of the online and the offline worlds in making these decisions and trying to come to what’s the right decision for your company or what’s the right product or service.

We just found increased reliance on online to do that research. When we say "increased reliance," we're probably talking 10 percentage points up over the three years. So, if 65 percent of the people were doing it in 2004, 75 percent of the people are doing it now. That’s primarily where we saw the trends going.

When we looked at the different phases of the buying cycle, it starts with awareness. You become aware that you need something. There was a high percentage of people -- in the high 60-percent range -- who said, "Once I become aware that I need something, the first place I'm going to go is the search engine to start looking for it." A lot of that traffic is going to end up on Google. It was the overwhelming choice among general search engines for B2B buyers.

But, as you move through the process, you start doing what we call a "landscape search." The first search is to get the lay of the land to figure out the information sites that have the information you are looking for. Who are the main vendors playing in this space? Where are the best bets to go and get more information to help make this purchase decision?

So, those first searches tend to be fairly generic -- shorter key phrases -- just to get the lay of the land to figure out where to go. As you progress, search tends to become more of a navigational shortcut, and we’ve seen this activity increase over the last two to three years. Increasingly, we're using search engines to get us from point A to point B online.

We also wanted to get a retroactive view of a successful transaction. So, in the second part of the survey, we asked them to recall a transaction they had made in the past 12 months. We wanted to see whether that initial search led to a successful purchase down the road, and, at the end of the road, how the different factors influenced them. So, we actually approached them from a couple of different angles.

Now, 85 percent of these people say they're using online search for some aspect of this purchasing process. It strikes me that this involves trillions of dollars worth of goods. These are big companies and, in some cases, buying lots of goods at over a hundred thousand dollars a whack. Do you concur that we're talking about trillions of dollars of B2B goods now being impacted significantly by the search process?

Absolutely. The importance of this is maybe the most mind-numbing fact to contemplate. Traditionally, the B2B space has been a little slow to move into the search arena. Traditionally, in the search arena, the big advertisers tend to be travel or financial products. B2B is just starting to understand how integral search is to all this activity. When you think of the nature of the B2B purchase, risk avoidance is a huge issue. You want to make sure that whatever decisions you make are well-researched and well-considered purchases. That naturally leads to a lot of online interaction.

The business information search is a primary factor driving [ZoomInfo's] growth. Our company right now is growing on two fronts. One is our traditional paid-search model, where we have subscription services focused on people information that is targeted at salespeople and recruiters as a source for candidates and prospects.

The more rapidly growing piece of our business is the advertising-driven business information search engine, which I think is a really interesting trend related to the concept you guys were just talking about. Not only does the B2B advertiser spend lots of money today trying to reach out, but the B2B searcher has new tools, services, and capabilities that provide a richer, better, more efficient search than they’ve had through the traditional search engines.

Everybody needs to be focused on search. I can’t see an exception. You mentioned the percentage that said they would go online. We segmented out the group that didn’t indicate they go online to see what was unique about them. The only thing unique about them was their age. They tended to be older buyers and tended to be with smaller organizations, where the CEO was more actively involved in the purchase decision. That was really the only variants we saw. If it’s a generational thing, then obviously that percentage is going to get smaller every year.

... From a vertical business information search perspective, that we’re really in the first inning here. A lot of interesting trends and enhancements are going to be coming down the road. One in particular that may have an influence in the next year or two is the community aspect within the search. ... I think that you’ll start to see a marriage of, not only B2B search, but also online community and a factoring into that whole process. Then, who knows where we’ll go from there? ... The word of community.
Listen to the podcast. Or read a full transcript. Sponsor: ZoomInfo.