chatgpt总结:. .и
Eric Lehman predicts that deep machine learning (ML) will outdo Google's search relevance algorithms soon, potentially within months. He highlights the recent success of BERT in web answers as an example of ML's rapid advancement. Lehman urges his team to prepare for this shift, noting that competitors could also develop superior ML systems. He recalls how Google Translate had to adapt when ML transformed translation, suggesting that web search relevance could face a similar disruption. Despite current systems' complexity, Lehman believes traditional methods will eventually fall behind as ML progresses, and he considers this shift to be almost inevitable. He encourages reflection on the future implications of this technological evolution.
原文:
On Wed, Dec 26, 2018 at 4:48 PM, Eric Lehman wrote:-baidu 1point3acres
I'd like to offer a thought for contemplation over the break: Within the near future, a deep ML system will clearly outperform Google's 20-year accumulation of relevance algorithms for web search. Here, I'm just talking about relevance; that is, determining whether a document and query are talking about the same thing. There is a lot more to web ranking for which ML seems much less appropriate. But I think basic relevance is the major task in web ranking and probably "objective" enough to go after pretty effectively with ML. None of us can see the future, but my bet is that this is nearly certain to be true within 5 years and could be true even within 6 months. One problem after another that is similar in flavor to web ranking has fallen, and there is little reason to think that web ranking is somehow exceptional. Indeed, this holiday thought stems from recent advances in web answers, where deep ML (in the form of BERT) abruptly subsumed essentially all preceding work. ..
For the web answers team, the tidal wave of deep ML that arrived in the last few weeks was a complete shock. With this warning, we should not allow ourselves to be caught off-guard again; rather, we should start thinking through the implications now. And now is really the time, because in the new year I expect a lot of web ranking engineers to reflect on BERT and start thinking along these same lines.
. .и
One consideration is that such a deep ML system could well be developed outside of Google—at Microsoft, Baidu, Yandex, Amazon, Apple, or even a startup. My impression is that the Translate team experienced this. Deep ML reset the translation game; past advantages were sort of wiped out. Fortunately, Google's huge investment in deep ML largely paid off, and we excelled in this new game. Nevertheless, our new ML-based translator was still beaten on benchmarks by a small startup. The risk that Google could similarly be beaten in relevance by another company is highlighted by a startling conclusion from BERT: huge amounts of user feedback can be largely replaced by unsupervised learning from raw text. That could have heavy implications for Google.
.--Relevance in web search may not fall quickly to deep ML, because we rely on memorization systems that are much larger than any current ML model and capture a ton of seemingly-crucial knowledge about language and the world. And there are lots of performance challenges, specialized considerations, etc. Still, my guess is that the advantages of our current approach will eventually crumble; ML is advancing very fast, and traditional techniques are not.
. Χ
I don't know how others think about this. Maybe this prospect was already obvious to you. Or you might think this view of the future is just wrong. Personally, I'm inclined to think that this future is near-inevitable, but—despite that—I hadn't taken the next step of thinking through implications.