This is a place devoted to giving you deeper insight
into the news, trends, people and technology behind Bing.
As I mentioned in my original post there has never been a more exciting or challenging time to be in the search space. The core to a great search engine has been and will always remain the same: delivering relevant, comprehensive and unbiased results that people can trust. We use thousands of signals from queries to documents and user feedback to determine the best search results and in turn make hundreds of improvements to our features every year, from small tweaks to core algorithm updates. In this series, you have heard from our Chief Scientist Dr. Jan Pedersen who summarized our efforts on whole page relevance, and Dr. Richard Qian who covered the techniques we employ to reduce junk links. More recently, Dr. William Ramsey described how we reduce defects in links and related searches and Dr. Kieran McDonald’s post examined answers quality utilizing specialized classifiers.
These are just a few examples of the myriad algorithm changes that we’ve developed to enhance Bing over the years. With all of these changes we’ve made some great progress in Bing search quality for our customers. You will notice we have released a fun, non-scientific tool for customers to see for themselves how far we’ve come. And while we know we still have lots of work to do, we think it’s long past time in our industry for a conversation on search quality. This conversation is what we hope to start with the Bing It On Challenge, and continue in the future to make sure we are delivering the quality experiences our customers deserve.
In this post I hope to give you a sense of what Bing’s journey has been like over the past couple of years in terms of our efforts to rapidly improve our search quality against a tough competitor.
How did we get here?
There isn’t a single answer to that question as there wasn’t just one silver bullet. We realized early on that experimentation is the key to improving relevance. We don’t know which feature ideas and techniques work until we actually try them out, and many ideas end up failing. So in order to have enough good ideas that succeed, we needed to build an experimentation funnel that enabled us to try out lots and lots of ideas to begin with. This was made more challenging by the fact that as a relatively new entrant to the search space, we had less traffic which made learning and experimentation harder.
So we invested in building a sophisticated experimentation system in the cloud which made it very easy for our engineers to try out new relevance techniques and train new machine learning models. This increased experimentation capacity allowed our engineers to submit hundreds of experiments in parallel and try out their ideas. This was a key breakthrough in scaling out relevance experimentation and enabling a large number of people on the team to effectively contribute in improving relevance and quality.
As experimentation volumes increased, we also realized that anything we could do to speed up this inner loop of experimentation would have a huge multiplier effect on how many more experiments we could run, and how quickly we could learn and implement new features to improve Bing’s quality.
Relevance experimentation at Bing involves training machine-learned models on large amount of training data using thousands of features. In the early years, our models were based on neural networks. But as the amount of training data, number of features and the complexity of our models increased, the inner loop of experimentation slowed down significantly. At one point, it took us several days to finish just one experiment end-to-end. We knew we needed to do something.
To overcome this challenge, we turned to our deep partnership with MSR to develop a technology we call Fastrank. FastRank is based on boosted decision trees which are much faster to train and thus attractive for relevance experimentation.. But there was skepticism on whether the quality of ranking produced by decision trees could match that of neural networks. Our colleagues at MSR took on this hard problem and developed new optimization algorithms that allowed us to not only match the quality of neural nets, but also train more than an order of magnitude faster.
This allowed us to once again make our inner experimentation loop fast, and really accelerate our pace of learning and improving ranking. In the end, our relevance improvements have come from trying out tens of thousands of ideas over the years, and shipping hundreds of those that have been successful.
When we previewed our side-by-side test results with people outside the company, I was often asked how we were able to make these gains with presumably less data than the other guys. While there are too many variables to give a fully scientific explanation, I would say our long-term commitment and investment in machine learning for relevance has enabled us to steadily scale out relevance experimentation and make rapid progress.
Of course, as we all know, relevance is subjective and queries are dynamic and always changing. But we feel confident that it’s time for customers to come give us a look, and for a conversation on searching quality to occur in our industry. This is why we’re inviting customers to try the Bing It On Challenge so that you have a fun way to compare side-by-side and come to your own conclusion. I encourage you to try it now at www.bing.com .
As always, we want to hear from you so please join the conversation.
On behalf of the Bing team
- Dr. Harry Shum, Corporate Vice President, Bing R&D
Tried it, google won all 15 rounds.
Why doesn't this site render correctly in IE9?
@shwa293 Thanks for taking the challenge. What issues are you experiencing on the site?
@schwa293 @Bing Search Blog
I'm using IE9. There's nothing wrong with the site. I don't see anything rendering incorrectly.
I have a question. When searching for terms like gold price, we do not see any videos on google/bing, but yet we see the videos on the bingiton challenge.
Hmmm, maybe i wasn't clear i was talking about this site. You know, the one you're reading this on right now.
I took the challenge twice:
First result was a 3 vs. 2 win for Google. The next result was a 4 vs. 1 win for Bing. I'd say Bing won overall.
One thing that makes the test not blind to the trained eye is the "X number of people +1'd this page" under Google results. To the trained eye, the tests aren't blind.
Great marketing idea and a good way to get the word out about this excellent search engine.
Thanks for blogging about this, Dr. Shum.
Thank you Bing and Dr. Shum for an interesting side-by-side comparison. My result was 3-Bing 2-draw. I have been watching Bing evolve, and I do believe that it continues proving to be the prefered choice. For several months I have been using Bing almost exclusively, my only using Google to check search ranks.
The item that I look for in search engine results is the quality of the sites listed. Of the keywords that I typically search for, Bing is far superior to all other search engines.
I do web design and SEO for a living. I have been keeping an eye on search engine ranking methods for about ten years, but I care more for my sites' content than search engine ranking. Some of my pages took over 300 hours to create and are based on years of research. I am pleased to find the pages ranked well on most all search engines, but the item of most interest is that Bing appears to recognize quality content better than other search engines.
We cannot please everyone, but I believe that Bing is doing very well and will hopefully become the most popular search engine in the near future. My sites' rankings are similar on Bing and Google, but I am more proud of my Bing ranking; it has a bit of class. :)
Bing - nice, very nice! I started using Bing a few years ago and occasionally go back a forth... But because of the quality of the results, I'm using Bing more and more: I just seem to find what I'm looking for better with Bing (and it's more fun and engaging).
Thank you for building a far better search engine.
Why are there so many duplicate posts here (not even counting the spam)?
Anyway, I got 3 Google and 2 Draw. But main problem I had was coming up with search terms when I don't really have any desire to search for something.
Took the challenge and google won twice but I know it's because that's what I've been use to and got really nervous and didn't take my time to compare. Sorry Bing but you still win hands down....
© 2013 Microsoft