CS 4476-A Computer Vision

Fall 2024, TR 11:00 to 12:15, Clough 144
Instructor: James Hays

TAs: Mengyu Yang, Jim James (head TAs), Akshath Shvetang Anna, Borun Song, Eric Zhang, Jinghao Miao, Kai Wang, Kristen Pereira, Meena Nagarajan, Qihang Hu, Sriram Goel, Zhenyu Wu

Computer Vision, art by kirkh.deviantart.com

Course Description

This course provides an introduction to computer vision including fundamentals of image formation, camera imaging geometry, feature detection and matching, stereo, motion estimation, convolutional networks, image classification, segmentation, object detection, transformers, and 3D computer vision. We'll explore methods for depth recovery from stereo images, camera calibration, automated alignment, tracking, boundary detection, and recognition. We'll use both classical machine learning and deep learning to approach these problems. The focus of the course is to develop the intuitions and mathematics of the methods in lecture, and then to implement substantial projects that resemble contemporary approaches to computer vision.

Learning Objectives

Upon completion of this course, students should be able to:

Prerequisites

No prior experience with computer vision is assumed, although previous knowledge of visual computing or signal processing will be helpful. The following skills are necessary for this class:

Compute Requirements

The institute has minimum compute requirements for all students that you are expected to meet. The projects are somewhat compute intensive, though, so a faster machine will let you iterate more quickly. The institute requirements say nothing about GPUs, unfortunately. The deep learning projects can benefit from (but will not require) a local GPU. Since we cannot rely on students having GPUs, projects will rely on cloud services such as Google Colab.

Grading

Your final grade will be made up from
Final grades will be A/B/C etc for 90.00%+, 80.00%+, 70.00%+, etc.

You will lose 10% each day for late projects. However, you have six "late days" for the whole course. That is to say, the first 24 hours after the due date and time counts as 1 day, up to 48 hours is two and 72 for the third late day. A late day cannot be split among projects (e.g. half a late day for project 1, and half a late day for project 2). They are used in integer amounts. There is no grace period for project handins. 5 minutes after the deadline is a full day late.

If you are taking this course, the expectation is that you have set aside the considerable amount of time needed to get your projects done in a timely manner. These late days are intended to cover unexpected clustering of due dates, travel commitments, interviews, hackathons, computer problems, extracurricular commitments, etc. Don't ask for extensions to due dates because we are already giving you a pool of late days to manage yourself. If you are seriously ill and need more time on projects beyond what late days can cover, you should submit documentation to the Dean of Students office and they will reach out to us.

Academic Integrity

Academic dishonesty will not be tolerated. This includes cheating, lying about course matters, plagiarism, code copying, or helping others commit a violation of the Honor Code. Plagiarism includes reproducing the words of others without both the use of quotation marks and citation. Students are reminded of the obligations and expectations associated with the Georgia Tech Academic Honor Code. We will use tools to find code sharing in projects.

You are expected to implement the core components of each project on your own, but the extra credit opportunties often build on third party data sets or code. That's fine. Feel free to include results built on other software, as long as your handin cites the toolboxes you build on.

You should not view, edit, or copy anyone else's code. You should not publicly post code to Canvas, except for starter code or helper code that isn't related to the core project.

More detailed guidance about what is and is not allowed in terms of collaboration with other students:
Allowed

Not allowed

Getting Help

The projects for this course are substantial and it will be common for students to need clarifications or guidance. Some of the explanation for projects will come from the lecture materials. Some guidance come from the project handout itself. If you are stumped or blocked, you can seek help from other students (within the parameters outlined above), you can ask for help on Piazza (but don't post any code unless your question is private), or you can attend TA office hours. Do not wait until the last few days of a project to post Piazza questions or attend office hours, because the TAs might be swamped.

The TAs job is not to debug your code. Do not dump entire source code files onto Piazza asking TAs to fix a bug. Do not hand your laptop to a TA at office hours and ask them to fix your code. The TAs primary goal is to make sure you understand the computer vision concepts necessary to complete the projects. Sometimes there will need to be discussions or troubleshooting at the code level, but don't be surprised if TAs try to minimize such interactions. If you want help from TAs, you should be ready to present the debugging that you've already done, e.g. "On this type of input, here's a visualization of the output that I'm getting".

Over the years, we have refined the projects and added unit tests and autograders so that students can more easily debug their own projects. Unfortunately, this sometimes leads to lazy development patterns. Students will express frustration that "my code passes the unit tests but the autograder gives errors" and expect the TAs to fix this situation. Keep in mind that the unit tests are necessary but not sufficient checks for code correctness. The unit tests help catch some commons errors, but it is expected that code might pass unit tests and fail the autograder if you did not implement something correctly. Even the autograder cannot catch all problems for all projects, because it can be difficult to check the correctness for certain computer vision algorithms. Your code might pass unit tests, get full marks from the autograder, but still have subtle errors that manifest elsewhere in the project. You will have more success in this class if you make an effort to understand the computer vision concepts being taught and not simply optimize for the autograder.

Learning Accommodations

If needed, we will make classroom accommodations for students with documented disabilities. These accommodations must be arranged in advance and in accordance with the office of disability services. (disabilityservices.gatech.edu).

Contact Info and Office Hours:

If possible, please use Piazza to ask questions and seek clarifications before emailing the instructor or staff. Office Hours

Projects (available on Canvas)

Optional Project 0: Test environment setup.
Project 1: Convolution and Hybrid images.
Project 2: SIFT Local Feature Matching and RANSAC.
Project 3: Recognition with deep learning.
Project 4: Semantic Segmentation.
Project 5: Classiflying Point Clouds with PointNet.
Possible Extra Credit Project 6: TBD.
All starter code and projects will be in Python with the use of various third party libraries. We will support MacOS and Windows. For MacOS, we are transitioning away from supporting Intel Processors. We will not support Linux (although things may just work out of the box). The course does not teach python and assumes you have enough familiarity with procedural programming languages to complete the projects.

Textbook

Readings will be assigned in "Computer Vision: Algorithms and Applications, 2nd edition" by Richard Szeliski. The book is available for purchase, but is also free to download.

Syllabus (tentative)

Class Date Topic Slides Reading
Tue, Aug 20 Introduction to computer vision pptx, pdf Szeliski 1
Image Formation and Filtering (Szeliski chapters 2 and 3)
Thu, Aug 22 Camera Projection and Image Filtering pptx, pdf Szeliski 2.1, especially 2.1.4
Tue, Aug 27 Thinking in Frequency pptx, pdf Szeliski 3.2 and 3.4
Thu, Aug 29 Thinking in Frequency, continued. pptx, pdf
Tue, Sep 3 Light, Cameras, Eyes, and Color. pptx, pdf Szeliski 2.2 and 2.3
Feature Detection and Matching
Thu, Sep 5 Interest point intro pptx, pdf Szeliski 7.1.1 and 7.1.2
Tue, Sep 10 Local image features, SIFT pptx, pdf Szeliski 7.1.3
Thu, Sep 12 Model fitting, Hough Transform pptx, pdf Szeliski 7.4.2 and 2.1
Tue, Sep 17 RANSAC, ICP, and geometric transformations pptx, pdf Szeliski 8.1 and 2.1
Multiple Views and Motion
Thu, Sep 19 Camera Calibration, Stereo intro pptx, pdf Szeliski 12 and 11.2.1
Tue, Sep 24 Guest lecture: Akshay Krishnan
Thu, Sep 26 Epipolar Geometry pptx, pdf Szeliski 11.3
Tue, Oct 1 Dense Stereo Correspondence pptx, pdf Szeliski 12
Thu, Oct 3 Quiz 1
Recognition
Tue, Oct 8 Machine learning crash course pptx, pdf Szeliski 5.1, 5.2, 5.3
Thu, Oct 10 Neural Networks and Convolutional Networks pptx, pdf Szeliski 5.3 and 5.4
Tue, Oct 15 No lecture, Institute holiday
Thu, Oct 17 Network Visualization and Quiz 1 recap pptx, pdf
Tue, Oct 22 Recognition techniques, old and new pptx, pdf Szeliski 6.2.1
Thu, Oct 24 Big Data and Geolocalization pptx, pdf Szeliski 6.3
Tue, Oct 29 Crowdsourcing and ResNet pptx, pdf
Thu, Oct 31 Semantic Segmentation pptx, pdf Szeliski 5.5
Tue, Nov 5 Self-Supervised Learning and Colorization pptx, pdf
Thu, Nov 7 Deep Object Detection and Structured Output from Deep Networks pptx, pdf Szeliski 5.4.7
Tue, Nov 12 3D Point Processing pptx, pdf
Thu, Nov 14 Transformer architectures
Tue, Nov 19 Generative Models - GANs and Diffusion
Thu, Nov 21 NeRF and Gaussian Splatting
Tue, Nov 26 Quiz 2
Thu, Nov 28 No lecture, Institute holiday
Tue, Dec 3 No lecture, work on project
Final Exam Period Not used. No final exam

Acknowledgements

The materials from this class rely significantly on slides prepared by other instructors, especially Derek Hoiem and Svetlana Lazebnik. Each slide set and assignment contains acknowledgements. Feel free to use these slides for academic or research purposes, but please maintain all acknowledgements.