Measuring Text Similarity Using BERT
This article was published as a part of the Data Science Blogathon BERT is too kind — so this article will be touching on BERT and sequence relationships! Abstract A significant portion of NLP relies on the connection in highly-dimensional spaces. Typically an NLP processing will take any text, prepare it to generate a tremendous vector/array rendering said text — then make certain transformations. It’s a highly-dimensional charm. At an exceptional level, there’s not much extra to it. We require to […]
Read more