With such a variety of normalization techniques how do multinational search engines such as google approach to the task? Do they use several different normalization algorithms for each language, or do they have only one very complex algorithm. If it’s different algorithms for different languages, then how do they cope with such documents that have instances of a different languages, for example language learning sites, where there are at least two different languages?

No comments:
Post a Comment