Difference between revisions of "BSc: Introduction To Computer Vision"
R.sirgalina (talk | contribs) |
|||
Line 1: | Line 1: | ||
+ | |||
= Introduction to Computer Vision = |
= Introduction to Computer Vision = |
||
+ | * '''Course name''': Introduction to Computer Vision |
||
+ | * '''Code discipline''': XXX |
||
+ | * '''Subject area''': |
||
+ | == Short Description == |
||
− | * <span>'''Course name:'''</span> Computer Vision |
||
+ | This course covers the following concepts: Computer vision techniques; Classical and deep learning models. |
||
− | * <span>'''Course number:'''</span> XXX |
||
− | == |
+ | == Prerequisites == |
− | === |
+ | === Prerequisite subjects === |
− | * Computer vision techniques |
||
− | * Classical and deep learning models |
||
+ | === Prerequisite topics === |
||
− | === What is the purpose of this course? === |
||
− | This course provides an intensive treatment of a cross-section of the key elements of computer vision, with an emphasis on implementing them in modern programming environments, and using them to solve real-world problems. The course will begin with the fundamentals of image processing and image filtering, but will quickly build to cover more advanced topics, including image segmentation, object detection and recognition, face detection, content-based image retrieval, artificial neural networks, convolutional neural networks, generative adversarial networks and much more. A key focus of the course is on providing students with not only theory but also hands-on practice of building their computer vision applications. |
||
− | == Course |
+ | == Course Topics == |
+ | {| class="wikitable" |
||
+ | |+ Course Sections and Topics |
||
+ | |- |
||
+ | ! Section !! Topics within the section |
||
+ | |- |
||
+ | | Image Acquisition and Basic Image Processing || |
||
+ | # Computer vision in action |
||
+ | # The Human Vision System |
||
+ | # Optical Illusions |
||
+ | # Sampling and Quantization |
||
+ | # Image Representation |
||
+ | # Colour Spaces |
||
+ | |- |
||
+ | | Image Filtering and Binary Vision || |
||
+ | # Image noise |
||
+ | # Convolutions and kernels |
||
+ | # Smoothing and blurring |
||
+ | # Thresholding and histograms |
||
+ | # Morphological operations |
||
+ | # Gradients and Edge detection |
||
+ | |- |
||
+ | | Feature Extractors and Descriptors || |
||
+ | # Histogram of Gradients (HoG) |
||
+ | # Scale-invariant feature transform (SIFT) |
||
+ | # Harris corner detector |
||
+ | # Template matching |
||
+ | # Bag of visual words |
||
+ | # Face Detection and Recognition (Viola Johns) |
||
+ | |- |
||
+ | | Deep learning models for computer vision || |
||
+ | # You Only Look Once: Unified, Real-Time Object Detection (YOLO) |
||
+ | # Generative Adversarial Networks (GAN) |
||
+ | # Fully Convolutional Networks (FCN) for semantic segmentation |
||
+ | # Multi Domain Network (MDNet) for object tracking |
||
+ | # Generic Object Tracking Using Regression Networks (GOTURN) for object tracking |
||
+ | |} |
||
+ | == Intended Learning Outcomes (ILOs) == |
||
− | === What |
+ | === What is the main purpose of this course? === |
+ | This course provides an intensive treatment of a cross-section of the key elements of computer vision, with an emphasis on implementing them in modern programming environments, and using them to solve real-world problems. The course will begin with the fundamentals of image processing and image filtering, but will quickly build to cover more advanced topics, including image segmentation, object detection and recognition, face detection, content-based image retrieval, artificial neural networks, convolutional neural networks, generative adversarial networks and much more. A key focus of the course is on providing students with not only theory but also hands-on practice of building their computer vision applications. |
||
+ | === ILOs defined at three levels === |
||
− | By the end of the course, the students should be able to process the video |
||
+ | ==== Level 1: What concepts should a student know/remember/explain? ==== |
||
+ | By the end of the course, the students should be able to ... |
||
* Significant exposure to real-world implementations |
* Significant exposure to real-world implementations |
||
* To develop research interest in the theory and application of computer vision |
* To develop research interest in the theory and application of computer vision |
||
− | === What should a student be able to |
+ | ==== Level 2: What basic practical skills should a student be able to perform? ==== |
+ | By the end of the course, the students should be able to ... |
||
− | |||
− | By the end of the course, the students should be able to choose the correct computer vision model. |
||
− | |||
* Suitability of different computer vision models in different scenarios |
* Suitability of different computer vision models in different scenarios |
||
* Ability to choose the right model for the given task |
* Ability to choose the right model for the given task |
||
− | === What should a student be able to apply |
+ | ==== Level 3: What complex comprehensive skills should a student be able to apply in real-life scenarios? ==== |
+ | By the end of the course, the students should be able to ... |
||
− | |||
− | By the end of the course, the students should be able to deploy and developed models. |
||
− | |||
* Hands on experience to implement different models to know inside behavior |
* Hands on experience to implement different models to know inside behavior |
||
* Sufficient exposure to train and deploy model for the given task |
* Sufficient exposure to train and deploy model for the given task |
||
− | * Fine tune the deployed model in the real-world settings |
+ | * Fine tune the deployed model in the real-world settings |
+ | == Grading == |
||
− | |||
− | === Course evaluation === |
||
+ | === Course grading range === |
||
− | {| |
||
+ | {| class="wikitable" |
||
− | |+ Course grade breakdown |
||
+ | |+ |
||
− | ! |
||
− | ! |
||
− | !align="center"| '''Proposed points''' |
||
|- |
|- |
||
+ | ! Grade !! Range !! Description of performance |
||
− | | Labs/seminar classes |
||
− | | 20 |
||
− | |align="center"| 20 |
||
|- |
|- |
||
+ | | A. Excellent || 90-100 || - |
||
− | | Interim performance assessment |
||
− | | 30 |
||
− | |align="center"| 50 |
||
|- |
|- |
||
+ | | B. Good || 75-89 || - |
||
− | | Exams |
||
− | | |
+ | |- |
+ | | C. Satisfactory || 60-74 || - |
||
− | |align="center"| 30 |
||
+ | |- |
||
+ | | D. Poor || 0-59 || - |
||
|} |
|} |
||
+ | === Course activities and grading breakdown === |
||
− | Labs/seminar classes points are divided into Quizzes and lab participation. Quizzes is 10% and two worst will be dropped. Lab homework is 10%. Interim performance assessment contains project with 20% weightage of the whole course. Exams will be conducted in theory and practical mode. |
||
+ | {| class="wikitable" |
||
− | |||
+ | |+ |
||
− | Late Submission Policy will be strictly applied in this course. Late submission automatically lowered one grade. If a personal emergency should arise that affects your ability, you must contact the course instructor BEFORE the deadline to get a “Special Late Submission Approval" from the course instructor. No “Special Late Submission Approval” will be granted after the deadline. There are no makeup quizzes. |
||
− | |||
− | === Grades range === |
||
− | |||
− | The grade ranges are defined in Table. [[#cvgrade|1]]. |
||
− | |||
− | <div id="cvgrade"> |
||
− | |||
− | {| |
||
− | |+ Course grading range |
||
− | ! |
||
− | ! |
||
− | !align="center"| '''Proposed range''' |
||
|- |
|- |
||
+ | ! Activity Type !! Percentage of the overall course grade |
||
− | | A. Excellent |
||
− | | 90-100 |
||
− | |align="center"| 90-100 |
||
|- |
|- |
||
+ | | Labs/seminar classes || 20 |
||
− | | B. Good |
||
− | | 75-89 |
||
− | |align="center"| 75-89 |
||
|- |
|- |
||
+ | | Interim performance assessment || 50 |
||
− | | C. Satisfactory |
||
− | | 60-74 |
||
− | |align="center"| 60-74 |
||
|- |
|- |
||
− | | |
+ | | Exams || 30 |
− | | 0-59 |
||
− | |align="center"| 0-59 |
||
|} |
|} |
||
+ | |||
+ | === Recommendations for students on how to succeed in the course === |
||
+ | == Resources, literature and reference materials == |
||
− | </div> |
||
− | === Resources and reference material === |
||
+ | === Open access resources === |
||
* Handouts supplied by the instructor |
* Handouts supplied by the instructor |
||
* Materials from the internet and research papers shared by instructor |
* Materials from the internet and research papers shared by instructor |
||
− | * |
||
− | * |
||
− | == |
+ | === Closed access resources === |
− | The main sections of the course and approximate hour distribution between them is as follows: |
||
+ | === Software and tools used within the course === |
||
− | {| |
||
+ | |||
− | |+ Course Sections |
||
+ | = Teaching Methodology: Methods, techniques, & activities = |
||
− | !align="center"| '''Section''' |
||
+ | |||
− | ! '''Section Title''' |
||
− | + | == Activities and Teaching Methods == |
|
+ | {| class="wikitable" |
||
+ | |+ Activities within each section |
||
|- |
|- |
||
+ | ! Learning Activities !! Section 1 !! Section 2 !! Section 3 !! Section 4 |
||
− | |align="center"| 1 |
||
− | | Image Acquisition and Basic Image Processing |
||
− | |align="center"| 8 |
||
|- |
|- |
||
+ | | Development of individual parts of software product code || 1 || 1 || 1 || 1 |
||
− | |align="center"| 2 |
||
− | | Image Filtering and Binary Vision |
||
− | |align="center"| 8 |
||
|- |
|- |
||
+ | | Homework and group projects || 1 || 1 || 1 || 1 |
||
− | |align="center"| 3 |
||
− | | Feature Extractors and Descriptors |
||
− | |align="center"| 16 |
||
|- |
|- |
||
+ | | Midterm evaluation || 1 || 1 || 1 || 1 |
||
− | |align="center"| 4 |
||
+ | |- |
||
− | | Deep Learning models for computer vision |
||
+ | | Testing (written or computer based) || 1 || 1 || 1 || 1 |
||
− | |align="center"| 16 |
||
− | | |
+ | |- |
+ | | Discussions || 1 || 1 || 1 || 1 |
||
+ | |} |
||
+ | == Formative Assessment and Course Activities == |
||
− | === |
+ | === Ongoing performance assessment === |
− | |||
− | ==== Section title: ==== |
||
− | |||
− | Image Acquisition and Basic Image Processing |
||
− | |||
− | === Topics covered in this section: === |
||
− | |||
− | * Computer vision in action |
||
− | * The Human Vision System |
||
− | * Optical Illusions |
||
− | * Sampling and Quantization |
||
− | * Image Representation |
||
− | * Colour Spaces |
||
− | |||
− | === What forms of evaluation were used to test students’ performance in this section? === |
||
− | |||
− | <div class="tabular"> |
||
− | |||
− | <span>|a|c|</span> & '''Yes/No'''<br /> |
||
− | Development of individual parts of software product code & 1<br /> |
||
− | Homework and group projects & 1<br /> |
||
− | Midterm evaluation & 1<br /> |
||
− | Testing (written or computer based) & 1<br /> |
||
− | Reports & 0<br /> |
||
− | Essays & 0<br /> |
||
− | Oral polls & 0<br /> |
||
− | Discussions & 1<br /> |
||
− | |||
− | |||
− | |||
− | </div> |
||
− | === Typical questions for ongoing performance evaluation within this section === |
||
− | |||
− | # What are the color spaces and where it’s used? |
||
− | # What are the primary and secondary colors? |
||
− | # How image is formed into computers? |
||
− | # How you will convert the RGB to grayscale images |
||
− | |||
− | === Typical questions for seminar classes (labs) within this section === |
||
− | |||
− | # Loading and plotting the images in python environment |
||
− | # Convertion of different color spaces |
||
− | # How you find the skin in the images based on the color space models |
||
− | # how to find red eye dot in face using color space models |
||
− | |||
− | === Test questions for final assessment in this section === |
||
+ | ==== Section 1 ==== |
||
+ | {| class="wikitable" |
||
+ | |+ |
||
+ | |- |
||
+ | ! Activity Type !! Content !! Is Graded? |
||
+ | |- |
||
+ | | Question || What are the color spaces and where it’s used? || 1 |
||
+ | |- |
||
+ | | Question || What are the primary and secondary colors? || 1 |
||
+ | |- |
||
+ | | Question || How image is formed into computers? || 1 |
||
+ | |- |
||
+ | | Question || How you will convert the RGB to grayscale images || 1 |
||
+ | |- |
||
+ | | Question || Loading and plotting the images in python environment || 0 |
||
+ | |- |
||
+ | | Question || Convertion of different color spaces || 0 |
||
+ | |- |
||
+ | | Question || How you find the skin in the images based on the color space models || 0 |
||
+ | |- |
||
+ | | Question || how to find red eye dot in face using color space models || 0 |
||
+ | |} |
||
+ | ==== Section 2 ==== |
||
+ | {| class="wikitable" |
||
+ | |+ |
||
+ | |- |
||
+ | ! Activity Type !! Content !! Is Graded? |
||
+ | |- |
||
+ | | Question || What are the challenges to perform histogram task? || 1 |
||
+ | |- |
||
+ | | Question || Apply convolutional filter to calculate the response. || 1 |
||
+ | |- |
||
+ | | Question || What kind of parameters are required to apply different image filters? || 1 |
||
+ | |- |
||
+ | | Question || How you will compute the gradients of the image and its benefits? || 1 |
||
+ | |- |
||
+ | | Question || Implement Otsu Method || 0 |
||
+ | |- |
||
+ | | Question || Implement Sobel, Preweitt filters || 0 |
||
+ | |- |
||
+ | | Question || Implement Canny edge detector || 0 |
||
+ | |- |
||
+ | | Question || Perform analysis over the different filtering on the given images || 0 |
||
+ | |} |
||
+ | ==== Section 3 ==== |
||
+ | {| class="wikitable" |
||
+ | |+ |
||
+ | |- |
||
+ | ! Activity Type !! Content !! Is Graded? |
||
+ | |- |
||
+ | | Question || How feature extractor works over the given image? || 1 |
||
+ | |- |
||
+ | | Question || What is the difference between the feature extraction and descriptors? || 1 |
||
+ | |- |
||
+ | | Question || Explain the examples of descriptors and feature extractors. || 1 |
||
+ | |- |
||
+ | | Question || Write down the pros and cons of SIFT, HOG and Harris. || 1 |
||
+ | |- |
||
+ | | Question || Implement template matching algorithm || 0 |
||
+ | |- |
||
+ | | Question || Implement histogram of gradient using CV2 library || 0 |
||
+ | |- |
||
+ | | Question || Implement of SIFT for the given task || 0 |
||
+ | |- |
||
+ | | Question || Implement Harris corner detection || 0 |
||
+ | |- |
||
+ | | Question || Analysis of different extractors for the given task || 0 |
||
+ | |} |
||
+ | ==== Section 4 ==== |
||
+ | {| class="wikitable" |
||
+ | |+ |
||
+ | |- |
||
+ | ! Activity Type !! Content !! Is Graded? |
||
+ | |- |
||
+ | | Question || How classification task is different from detection task? || 1 |
||
+ | |- |
||
+ | | Question || Explain the transfer learning mechanism for object detection task. || 1 |
||
+ | |- |
||
+ | | Question || How many types of model exist for object tracking in videos. || 1 |
||
+ | |- |
||
+ | | Question || Write down the pros and cons of YOLO, FCN and MDNet. || 1 |
||
+ | |- |
||
+ | | Question || Implement YOLO using transfer learning mechanism || 0 |
||
+ | |- |
||
+ | | Question || Implement GAN for MNIST dataset || 0 |
||
+ | |- |
||
+ | | Question || Implement FCN and GOTURN || 0 |
||
+ | |- |
||
+ | | Question || Analysis of different models for the given task || 0 |
||
+ | |} |
||
+ | === Final assessment === |
||
+ | '''Section 1''' |
||
# How you can distinguish different color spaces? |
# How you can distinguish different color spaces? |
||
# Explain and provide the reason for the blind spot creation in human eye. |
# Explain and provide the reason for the blind spot creation in human eye. |
||
# In what scenarios computer vision is better than human vision? |
# In what scenarios computer vision is better than human vision? |
||
# Write down different robotic application areas where computer vision is applied successfully. |
# Write down different robotic application areas where computer vision is applied successfully. |
||
+ | '''Section 2''' |
||
− | |||
− | === Section 2 === |
||
− | |||
− | ==== Section title: ==== |
||
− | |||
− | Image Filtering and Binary Vision |
||
− | |||
− | === Topics covered in this section: === |
||
− | |||
− | * Image noise |
||
− | * Convolutions and kernels |
||
− | * Smoothing and blurring |
||
− | * Thresholding and histograms |
||
− | * Morphological operations |
||
− | * Gradients and Edge detection |
||
− | |||
− | === What forms of evaluation were used to test students’ performance in this section? === |
||
− | |||
− | <div class="tabular"> |
||
− | |||
− | <span>|a|c|</span> & '''Yes/No'''<br /> |
||
− | Development of individual parts of software product code & 1<br /> |
||
− | Homework and group projects & 1<br /> |
||
− | Midterm evaluation & 1<br /> |
||
− | Testing (written or computer based) & 1<br /> |
||
− | Reports & 0<br /> |
||
− | Essays & 0<br /> |
||
− | Oral polls & 0<br /> |
||
− | Discussions & 1<br /> |
||
− | |||
− | |||
− | |||
− | </div> |
||
− | === Typical questions for ongoing performance evaluation within this section === |
||
− | |||
− | # What are the challenges to perform histogram task? |
||
− | # Apply convolutional filter to calculate the response. |
||
− | # What kind of parameters are required to apply different image filters? |
||
− | # How you will compute the gradients of the image and its benefits? |
||
− | |||
− | === Typical questions for seminar classes (labs) within this section === |
||
− | |||
− | # Implement Otsu Method |
||
− | # Implement Sobel, Preweitt filters |
||
− | # Implement Canny edge detector |
||
− | # Perform analysis over the different filtering on the given images |
||
− | |||
− | === Test questions for final assessment in this section === |
||
− | |||
# Calculate the kernels for the given images |
# Calculate the kernels for the given images |
||
# Explain the difference between different filters |
# Explain the difference between different filters |
||
# What is image noise and how it contributes to make the computer vision task difficult? |
# What is image noise and how it contributes to make the computer vision task difficult? |
||
# Apply different combination of the filters to achieve the required output of the given image. |
# Apply different combination of the filters to achieve the required output of the given image. |
||
+ | '''Section 3''' |
||
− | |||
− | === Section 3 === |
||
− | |||
− | ==== Section title: ==== |
||
− | |||
− | Feature Extractors and Descriptors |
||
− | |||
− | ==== Topics covered in this section: ==== |
||
− | |||
− | * Histogram of Gradients (HoG) |
||
− | * Scale-invariant feature transform (SIFT) |
||
− | * Harris corner detector |
||
− | * Template matching |
||
− | * Bag of visual words |
||
− | * Face Detection and Recognition (Viola Johns) |
||
− | |||
− | === What forms of evaluation were used to test students’ performance in this section? === |
||
− | |||
− | <div class="tabular"> |
||
− | |||
− | <span>|a|c|</span> & '''Yes/No'''<br /> |
||
− | Development of individual parts of software product code & 1<br /> |
||
− | Homework and group projects & 1<br /> |
||
− | Midterm evaluation & 1<br /> |
||
− | Testing (written or computer based) & 1<br /> |
||
− | Reports & 0<br /> |
||
− | Essays & 0<br /> |
||
− | Oral polls & 0<br /> |
||
− | Discussions & 1<br /> |
||
− | |||
− | |||
− | |||
− | </div> |
||
− | === Typical questions for ongoing performance evaluation within this section === |
||
− | |||
− | # How feature extractor works over the given image? |
||
− | # What is the difference between the feature extraction and descriptors? |
||
− | # Explain the examples of descriptors and feature extractors. |
||
− | # Write down the pros and cons of SIFT, HOG and Harris. |
||
− | |||
− | ==== Typical questions for seminar classes (labs) within this section ==== |
||
− | |||
− | # Implement template matching algorithm |
||
− | # Implement histogram of gradient using CV2 library |
||
− | # Implement of SIFT for the given task |
||
− | # Implement Harris corner detection |
||
− | # Analysis of different extractors for the given task |
||
− | |||
− | ==== Test questions for final assessment in this section ==== |
||
− | |||
# How you distinguish different feature extractors and descriptors? |
# How you distinguish different feature extractors and descriptors? |
||
# What are the possible methods to detect the corners? |
# What are the possible methods to detect the corners? |
||
# How corners are useful to help the robotic vision task? |
# How corners are useful to help the robotic vision task? |
||
# How you will patch the different images to construct the map of the location? |
# How you will patch the different images to construct the map of the location? |
||
+ | '''Section 4''' |
||
+ | # What are the loss functions used in YOLO? |
||
+ | # What are the learnable parameters of FCN for semantic segmentation? |
||
+ | # How semantic segmentation is different from instance segmentation? |
||
+ | # Write the application areas for object tracking in robotics. |
||
− | === |
+ | === The retake exam === |
+ | '''Section 1''' |
||
− | + | '''Section 2''' |
|
+ | '''Section 3''' |
||
− | Deep learning models for computer vision |
||
+ | '''Section 4''' |
||
− | ==== Topics covered in this section: ==== |
||
− | |||
− | * You Only Look Once: Unified, Real-Time Object Detection (YOLO) |
||
− | * Generative Adversarial Networks (GAN) |
||
− | * Fully Convolutional Networks (FCN) for semantic segmentation |
||
− | * Multi Domain Network (MDNet) for object tracking |
||
− | * Generic Object Tracking Using Regression Networks (GOTURN) for object tracking |
||
− | |||
− | === What forms of evaluation were used to test students’ performance in this section? === |
||
− | |||
− | <div class="tabular"> |
||
− | |||
− | <span>|a|c|</span> & '''Yes/No'''<br /> |
||
− | Development of individual parts of software product code & 1<br /> |
||
− | Homework and group projects & 1<br /> |
||
− | Midterm evaluation & 1<br /> |
||
− | Testing (written or computer based) & 1<br /> |
||
− | Reports & 0<br /> |
||
− | Essays & 0<br /> |
||
− | Oral polls & 0<br /> |
||
− | Discussions & 1<br /> |
||
− | |||
− | |||
− | |||
− | </div> |
||
− | === Typical questions for ongoing performance evaluation within this section === |
||
− | |||
− | # How classification task is different from detection task? |
||
− | # Explain the transfer learning mechanism for object detection task. |
||
− | # How many types of model exist for object tracking in videos. |
||
− | # Write down the pros and cons of YOLO, FCN and MDNet. |
||
− | |||
− | ==== Typical questions for seminar classes (labs) within this section ==== |
||
− | |||
− | # Implement YOLO using transfer learning mechanism |
||
− | # Implement GAN for MNIST dataset |
||
− | # Implement FCN and GOTURN |
||
− | # Analysis of different models for the given task |
||
− | |||
− | ==== Test questions for final assessment in this section ==== |
||
− | |||
− | # What are the loss functions used in YOLO? |
||
− | # What are the learnable parameters of FCN for semantic segmentation? |
||
− | # How semantic segmentation is different from instance segmentation? |
||
− | # Write the application areas for object tracking in robotics. |
Revision as of 12:57, 12 July 2022
Introduction to Computer Vision
- Course name: Introduction to Computer Vision
- Code discipline: XXX
- Subject area:
Short Description
This course covers the following concepts: Computer vision techniques; Classical and deep learning models.
Prerequisites
Prerequisite subjects
Prerequisite topics
Course Topics
Section | Topics within the section |
---|---|
Image Acquisition and Basic Image Processing |
|
Image Filtering and Binary Vision |
|
Feature Extractors and Descriptors |
|
Deep learning models for computer vision |
|
Intended Learning Outcomes (ILOs)
What is the main purpose of this course?
This course provides an intensive treatment of a cross-section of the key elements of computer vision, with an emphasis on implementing them in modern programming environments, and using them to solve real-world problems. The course will begin with the fundamentals of image processing and image filtering, but will quickly build to cover more advanced topics, including image segmentation, object detection and recognition, face detection, content-based image retrieval, artificial neural networks, convolutional neural networks, generative adversarial networks and much more. A key focus of the course is on providing students with not only theory but also hands-on practice of building their computer vision applications.
ILOs defined at three levels
Level 1: What concepts should a student know/remember/explain?
By the end of the course, the students should be able to ...
- Significant exposure to real-world implementations
- To develop research interest in the theory and application of computer vision
Level 2: What basic practical skills should a student be able to perform?
By the end of the course, the students should be able to ...
- Suitability of different computer vision models in different scenarios
- Ability to choose the right model for the given task
Level 3: What complex comprehensive skills should a student be able to apply in real-life scenarios?
By the end of the course, the students should be able to ...
- Hands on experience to implement different models to know inside behavior
- Sufficient exposure to train and deploy model for the given task
- Fine tune the deployed model in the real-world settings
Grading
Course grading range
Grade | Range | Description of performance |
---|---|---|
A. Excellent | 90-100 | - |
B. Good | 75-89 | - |
C. Satisfactory | 60-74 | - |
D. Poor | 0-59 | - |
Course activities and grading breakdown
Activity Type | Percentage of the overall course grade |
---|---|
Labs/seminar classes | 20 |
Interim performance assessment | 50 |
Exams | 30 |
Recommendations for students on how to succeed in the course
Resources, literature and reference materials
Open access resources
- Handouts supplied by the instructor
- Materials from the internet and research papers shared by instructor
Closed access resources
Software and tools used within the course
Teaching Methodology: Methods, techniques, & activities
Activities and Teaching Methods
Learning Activities | Section 1 | Section 2 | Section 3 | Section 4 |
---|---|---|---|---|
Development of individual parts of software product code | 1 | 1 | 1 | 1 |
Homework and group projects | 1 | 1 | 1 | 1 |
Midterm evaluation | 1 | 1 | 1 | 1 |
Testing (written or computer based) | 1 | 1 | 1 | 1 |
Discussions | 1 | 1 | 1 | 1 |
Formative Assessment and Course Activities
Ongoing performance assessment
Section 1
Activity Type | Content | Is Graded? |
---|---|---|
Question | What are the color spaces and where it’s used? | 1 |
Question | What are the primary and secondary colors? | 1 |
Question | How image is formed into computers? | 1 |
Question | How you will convert the RGB to grayscale images | 1 |
Question | Loading and plotting the images in python environment | 0 |
Question | Convertion of different color spaces | 0 |
Question | How you find the skin in the images based on the color space models | 0 |
Question | how to find red eye dot in face using color space models | 0 |
Section 2
Activity Type | Content | Is Graded? |
---|---|---|
Question | What are the challenges to perform histogram task? | 1 |
Question | Apply convolutional filter to calculate the response. | 1 |
Question | What kind of parameters are required to apply different image filters? | 1 |
Question | How you will compute the gradients of the image and its benefits? | 1 |
Question | Implement Otsu Method | 0 |
Question | Implement Sobel, Preweitt filters | 0 |
Question | Implement Canny edge detector | 0 |
Question | Perform analysis over the different filtering on the given images | 0 |
Section 3
Activity Type | Content | Is Graded? |
---|---|---|
Question | How feature extractor works over the given image? | 1 |
Question | What is the difference between the feature extraction and descriptors? | 1 |
Question | Explain the examples of descriptors and feature extractors. | 1 |
Question | Write down the pros and cons of SIFT, HOG and Harris. | 1 |
Question | Implement template matching algorithm | 0 |
Question | Implement histogram of gradient using CV2 library | 0 |
Question | Implement of SIFT for the given task | 0 |
Question | Implement Harris corner detection | 0 |
Question | Analysis of different extractors for the given task | 0 |
Section 4
Activity Type | Content | Is Graded? |
---|---|---|
Question | How classification task is different from detection task? | 1 |
Question | Explain the transfer learning mechanism for object detection task. | 1 |
Question | How many types of model exist for object tracking in videos. | 1 |
Question | Write down the pros and cons of YOLO, FCN and MDNet. | 1 |
Question | Implement YOLO using transfer learning mechanism | 0 |
Question | Implement GAN for MNIST dataset | 0 |
Question | Implement FCN and GOTURN | 0 |
Question | Analysis of different models for the given task | 0 |
Final assessment
Section 1
- How you can distinguish different color spaces?
- Explain and provide the reason for the blind spot creation in human eye.
- In what scenarios computer vision is better than human vision?
- Write down different robotic application areas where computer vision is applied successfully.
Section 2
- Calculate the kernels for the given images
- Explain the difference between different filters
- What is image noise and how it contributes to make the computer vision task difficult?
- Apply different combination of the filters to achieve the required output of the given image.
Section 3
- How you distinguish different feature extractors and descriptors?
- What are the possible methods to detect the corners?
- How corners are useful to help the robotic vision task?
- How you will patch the different images to construct the map of the location?
Section 4
- What are the loss functions used in YOLO?
- What are the learnable parameters of FCN for semantic segmentation?
- How semantic segmentation is different from instance segmentation?
- Write the application areas for object tracking in robotics.
The retake exam
Section 1
Section 2
Section 3
Section 4