Last week, Microsoft released a new Bing image search experience designed to help customers be inspired, learn more, and do more with image search. Today, Bing team provided a deeper look at how they built the image search graph that allows us to automatically derive useful descriptions, captions and actions for images in the new search experience.
Here are some interesting info from their blog post,
- Our data shows that 59% of images have at least one duplicate somewhere on the web. As illustrated in the chart below, images are often duplicated hundreds – even thousands of times.
- Fortunately, modern Image Understanding techniques create a big opportunity to bring all the metadata together for exact or near duplicate image content that is currently fragmented across many URLs. Bing seizes this opportunity by analyzing every image’s content and context and constructs an advanced graph clustering similar images and all their associated metadata together. This graph, which is pivoted on content within the image rather than URLs, is what we refer to as the Image Graph and is the foundational building block of our new experience.
- With the Image Graph, we are able to find deep links among collections. If one image appears in a few different collections, these collections are connected with each other.
- We leveraged the Image Graph to mine user interactions with its content clusters. We also looked at the web page context where images are hosted. With this data, we built an algorithm to extract the image’s primary focus. The algorithm has been applied to billions of images and the BRQ is preserved in the Image Graph. In Bing Image Search, the BRQ is displayed on the hover of search result as well as in the new image viewing experience. The BRQ is designed to be search-friendly, so you are one click away from learning more about the key concept for each image result.
Read about it in detail from the link below.