\relax \select@language{english} \@writefile{toc}{\select@language{english}} \@writefile{lof}{\select@language{english}} \@writefile{lot}{\select@language{english}} \@writefile{toc}{\contentsline {section}{\numberline {1}Documents and query representation}{1}} \@writefile{toc}{\contentsline {subsection}{\numberline {1.1}Term incidence matrix}{1}} \@writefile{lof}{\contentsline {figure}{\numberline {1}{\ignorespaces Term Incidence Matrix for the William Shakespeare plays}}{1}} \@writefile{toc}{\contentsline {subsection}{\numberline {1.2}Retrieval Models}{2}} \@writefile{lof}{\contentsline {figure}{\numberline {2}{\ignorespaces Basic Retrieval Model}}{2}} \@writefile{toc}{\contentsline {subsection}{\numberline {1.3}Bag of words representation}{2}} \@writefile{toc}{\contentsline {section}{\numberline {2}Retrieval Scoring Function, Matching}{3}} \@writefile{toc}{\contentsline {subsection}{\numberline {2.1}Boolean Retrieval}{3}} \@writefile{toc}{\contentsline {section}{\numberline {3}Vector Space Models}{4}} \@writefile{toc}{\contentsline {subsection}{\numberline {3.1}DLength Normalization}{5}} \@writefile{lof}{\contentsline {figure}{\numberline {3}{\ignorespaces Drawback of considering raw TF without normalization}}{5}} \@writefile{toc}{\contentsline {subsection}{\numberline {3.2}Okapi TF and other TF scaling values instead of TF}{5}} \@writefile{toc}{\contentsline {subsection}{\numberline {3.3}Similarity of vectors}{6}} \@writefile{lof}{\contentsline {figure}{\numberline {4}{\ignorespaces Dot Product Similarity Scores}}{6}} \@writefile{lof}{\contentsline {figure}{\numberline {5}{\ignorespaces Cosine Similarity Scores}}{7}} \@writefile{toc}{\contentsline {subsection}{\numberline {3.4}TF-IDF}{7}} \@writefile{lof}{\contentsline {figure}{\numberline {6}{\ignorespaces Various term score functions including tf-idf}}{8}} \@writefile{toc}{\contentsline {section}{\numberline {4}Term Concurrence}{8}} \@writefile{lof}{\contentsline {figure}{\numberline {7}{\ignorespaces Measures of Co-Occurrence}}{9}} \@writefile{toc}{\contentsline {section}{\numberline {5}BM25}{11}} \@writefile{toc}{\contentsline {section}{\numberline {6}Language Models}{12}} \@writefile{toc}{\contentsline {subsection}{\numberline {6.1} basics of probability likelihoods}{13}} \@writefile{toc}{\contentsline {subsection}{\numberline {6.2}Query Likelihood Model}{13}} \@writefile{toc}{\contentsline {subsection}{\numberline {6.3}Divergence Language Model}{13}} \@writefile{toc}{\contentsline {subsection}{\numberline {6.4}Implementation of Language Models}{13}} \@writefile{toc}{\contentsline {subsection}{\numberline {6.5}Estimating Probabilities}{13}} \@writefile{toc}{\contentsline {subsection}{\numberline {6.6}Smoothing}{13}} \@writefile{toc}{\contentsline {subsection}{\numberline {6.7}Language Models with Ngrams}{13}} \@writefile{toc}{\contentsline {section}{\numberline {7}Skipgram Minimum Span Model}{13}} \@writefile{toc}{\contentsline {section}{\numberline {8}Query Expansion, Relevance Feedback}{13}} \@writefile{toc}{\contentsline {section}{\numberline {9}Metasearch- Ranking Fusion}{13}} \@writefile{toc}{\contentsline {section}{\numberline {10}Evaluate with trec\_eval against a qrel file}{13}}