Week 1 OverviewHelp Center

Week 1

On this page:

Instructional Activities

Below is a list of the activities and assignments available to you this week. Click on the name of each activity for more detailed instructions.

Relevant Badges Activity Due Date* Estimated Time Required
  Week 1 Video Lectures Sunday, March 29 (Suggested) 3 hours
Programming Assignments Overview Sunday, March 29
(Suggested)
~1 hour
Week 1 Quiz Sunday, April 19 ~ 0.5 hour

* All deadlines are at 11:55 PM Central Time (time zone conversion) unless otherwise noted.

Time

This module will last 7 days and should take approximately 5 hours of dedicated time to complete, with its readings and assignments.

Goals and Objectives

After you actively engage in the learning experiences in this module, you should be able to:

  • Explain some basic concepts in natural language processing and text information access.
  • Explain why text retrieval is often defined as a ranking problem.
  • Explain how the vector space retrieval model works.
  • Explain what TF-IDF weighting is and why TF transformation and document length normalization is necessary for the design of an effective ranking function.

Key Phrases/Concepts

Keep your eyes open for the following key terms or phrases as you complete the readings and interact with the lectures. These topics will help you better understand the content in this module.

  • Part-of-speech tagging; syntactic analysis; semantic analysis; ambiguity
  • “Bag of words” representation
  • Push, pull, querying, browsing
  • Probability Ranking Principle
  • Relevance
  • Vector Space Model
  • Term Frequency (TF)
  • Document Frequency (DF); Inverse Document Frequency (IDF)
  • TF Transformation
  • Pivoted length normalization
  • Dot product
  • BM25

Guiding Questions

Develop your answers to the following guiding questions while watching the video lectures throughout the week.

  • What does a computer have to do in order to understand a natural language sentence?
  • What is ambiguity?
  • Why is natural language processing (NLP) difficult for computers?
  • What is bag-of-words representation? Why do modern search engines use this simple representation of text?
  • What are the two modes of text information access? Which mode does a Web search engine such as Google support?
  • When is browsing more useful than querying to help a user find relevant information?
  • Why is a text retrieval task defined as a ranking task?
  • What is a retrieval model?
  • What are the two assumptions made by the Probability Ranking Principle?
  • What is the Vector Space Retrieval Model? How does it work?
  • How do we define the dimensions of the Vector Space Model?
  • What are some different ways to place a document as a vector in the vector space?
  • What is Term Frequency (TF)?
  • What is TF Transformation?
  • What is Document Frequency (DF)?
  • What is Inverse Document Frequency (IDF)?
  • What is TF-IDF Weighting?
  • Why do we need to penalize long documents in text retrieval?
  • What is pivoted document length normalization?
  • What are the main ideas behind the retrieval function BM25?

Readings and Resources

The following readings are optional:

  • N. J. Belkin and W. B. Croft. "Information filtering and information retrieval: Two sides of the same coin?" Commun. ACM 35, 12 (Dec. 1992): 29-38.
  • A. Singhal, C. Buckley, and M. Mitra. "Pivoted document length normalization." In Proceedings of ACM SIGIR 1996.

Video Lectures

Video Lecture Lecture Notes Transcript Video Download SRT Caption File Forum
 1.1 Natural Language Processing(00:21:05)    
 
(35.5 MB)
   
 1.2 Text Access(00:09:24)    
 
(12.8 MB)
   
 1.3 Text Retrieval Problem(00:26:18)    
 
(36.7 MB)
   
 1.4 Overview of Text Retrieval Methods(00:10:10)    
 
(13.7 MB)
   
 1.5 Vector Space Model: Basic Idea(00:09:44)    
 
(13.0 MB)
   
 1.6 Vector Space Model: Instantiation(00:17:30)    
 
(23.1 MB)
   
 1.7 Vector Space Model: Improved Instantiation(00:16:52)    
 
(22.1 MB)
   
 1.8 TF Transformation (00:18:56)    
 
(12.7 MB)
   
 1.9 Doc Length Normalization(00:18:56)    
 
(25.6 MB)
   

Tips for Success

To do well this week, I recommend that you do the following:

  • Review the video lectures a number of times to gain a solid understanding of the key questions and concepts introduced this week.
  • When possible, provide tips and suggestions to your peers in this class. As a learning community, we can help each other learn and grow. One way of doing this is by helping to address the questions that your peers pose. By engaging with each other, we’ll all learn better.
  • It’s always a good idea to refer to the video lectures and reference them in your responses. When appropriate, critique the information presented.
  • Take notes while you watch the lectures for this week. By taking notes, you are interacting with the material and will find that it is easier to remember and to understand. With your notes, you’ll also find that it’s easier to complete your assignments. So, go ahead, do yourself a favor; take some notes!

Getting and Giving Help

You can get/give help via the following means:

  • Use the Learner Help Center to find information regarding specific technical problems. For example, technical problems would include error messages, difficulty submitting assignments, or problems with video playback. You can access the Help Center by clicking on theHelp Center link at the top right of any course page. If you cannot find an answer in the documentation, you can also report your problem to the Coursera staff by clicking on the Contact Us! link available on each topic's page within the Learner Help Center.
  • Use the Content Issues forum to report errors in lecture video content, assignment questions and answers, assignment grading, text and links on course pages, or the content of other course materials. University of Illinois staff and Community TAs will monitor this forum and respond to issues.

As a reminder, the instructor is not able to answer emails sent directly to his account. Rather, all questions should be reported as described above.

from: https://class.coursera.org/textretrieval-001/wiki/Week1Overview

coursera课程Text Retrieval and Search Engines之Week 1 Overview的更多相关文章

  1. coursera课程Text Retrieval and Search Engines之Week 2 Overview

    Week 2 OverviewHelp Center Week 2 On this page: Instructional Activities Time Goals and Objectives K ...

  2. coursera课程Text Retrieval and Search Engines之Week 3 Overview

    Week 3 OverviewHelp Center Week 3 On this page: Instructional Activities Time Goals and Objectives K ...

  3. coursera课程Text Retrieval and Search Engines之Week 4 Overview

    Week 4 OverviewHelp Center Week 4 On this page: Instructional Activities Time Goals and Objectives K ...

  4. 【Python学习笔记】Coursera课程《Using Databases with Python》 密歇根大学 Charles Severance——Week4 Many-to-Many Relationships in SQL课堂笔记

    Coursera课程<Using Databases with Python> 密歇根大学 Week4 Many-to-Many Relationships in SQL 15.8 Man ...

  5. 【Python学习笔记】Coursera课程《Using Python to Access Web Data》 密歇根大学 Charles Severance——Week6 JSON and the REST Architecture课堂笔记

    Coursera课程<Using Python to Access Web Data> 密歇根大学 Week6 JSON and the REST Architecture 13.5 Ja ...

  6. 【Python学习笔记】Coursera课程《Using Python to Access Web Data 》 密歇根大学 Charles Severance——Week2 Regular Expressions课堂笔记

    Coursera课程<Using Python to Access Web Data > 密歇根大学 Charles Severance Week2 Regular Expressions ...

  7. Coursera课程下载和存档计划[转载]

    上周三收到Coursera平台的群发邮件,大意是Coursera将在6月30号彻底关闭旧的课程平台,全面升级到新的课程平台上,一些旧的课程资源(课程视频.课程资料)将不再保存,如果你之前学习过相关的课 ...

  8. 【网页开发学习】Coursera课程《面向 Web 开发者的 HTML、CSS 与 Javascript》Week1课堂笔记

    Coursera课程<面向 Web 开发者的 HTML.CSS 与 Javascript> Johns Hopkins University Yaakov Chaikin Week1 In ...

  9. 【DeepLearning学习笔记】Coursera课程《Neural Networks and Deep Learning》——Week2 Neural Networks Basics课堂笔记

    Coursera课程<Neural Networks and Deep Learning> deeplearning.ai Week2 Neural Networks Basics 2.1 ...

随机推荐

  1. USACO 6.4 Wisconsin Squares

    Wisconsin Squares It's spring in Wisconsin and time to move the yearling calves to the yearling past ...

  2. WIN8+VS2013编写发布WCF、一(编写)、二(部署)、三(调用)

    原文://http://www.cnblogs.com/tntboom/p/4348483.html 引言:上学期因为写服务器用WCF,所以连查资料再瞎调试勉强成功了,但是这学期又到了用WCF的时候, ...

  3. linux 命令行对mysql数据进行备份和还原

    一.备份 1.进入mysql目录 cd /var/lib/mysql (进入mysql目录,根据安装情况会有差别) 2.备份 mysqldump -u root -p密码 数据库名 数据表名 > ...

  4. Kubernetes(k8s)集群部署(k8s企业级Docker容器集群管理)系列之自签TLS证书及Etcd集群部署(二)

    0.前言 整体架构目录:ASP.NET Core分布式项目实战-目录 k8s架构目录:Kubernetes(k8s)集群部署(k8s企业级Docker容器集群管理)系列目录 一.服务器设置 1.把每一 ...

  5. Redis 服务器命令

    1.BGREWRITEAOF 异步执行一个 AOF(AppendOnly File) 文件重写操作 2.BGSAVE 在后台异步保存当前数据库的数据到磁盘 3.CLIENT KILL [ip:port ...

  6. 每日踩坑 2018-06-19 AutoMapper简单性能测试

    想使用 AutoMapper 类库来做一些映射到 DTO 对象的操作 但既然类似这样的类库内部是用反射来实现的,那么会比较在意性能. 所以来简单测试一下性能. 关于测试结果呢 emmmm 我是比较吃惊 ...

  7. SPOJ7586 NUMOFPAL manacher算法

    题目大意: 求一个串中有多少个回文子串 这..... 妥妥的模板题吧.... 对所有的$r[i] / 2$进行求和即可,其中,$r[i]$为以$i$为中心的回文半径 $r[i] / 2$怎么来的,画下 ...

  8. java知识回顾

    一.构造方法能不能被继承 当然不能,1.构造方法是类的唯一入口 2.构造方法与类名相同 3.子类构造方法中隐式的调用了父类的构造方法 二.值传递和引用传递.不变类和可变类.直接赋值和浅拷贝和深拷贝   ...

  9. Codeforces Round #131 (Div. 1) A - Game

    A. Game time limit per test 1 second memory limit per test 256 megabytes input standard input output ...

  10. SMACH(五)----用户数据UserData类和重映射Remapper类的原理和例子

    用户数据UserData类和重映射Remapper类包含在smach中的user_data.py文件中实现,该博文主要介绍其原理和例子 UserData主要用于状态之间的数据传递,包括数据的输入inp ...