Lecture_14_object_detection_segmentation Project Proposal Instructions

User Manual:

Open the PDF directly: View PDF PDF.
Page Count: 8

DownloadLecture_14_object_detection_segmentation  Project Proposal Instructions
Open PDF In BrowserView PDF
Class project details
• Can work alone or in a group (up to 4 people), required effort will scale with # of people
• Select a “base” dataset (online, or from a list I’ll make)
• Simulate parameters of a physical (imaging) system with base dataset
• Train deep neural net with simulated dataset
• Report results

Class project details
• Can work alone or in a group (up to 4 people), required effort will scale with # of people
• Select a “base” dataset (online, or from a list I’ll make)
• Simulate parameters of a physical (imaging) system with base dataset
• Train deep neural net with simulated dataset
• Report results
What you’ll need to submit:
1) The project’s source code
2) A short research-style paper (3 pages minimum, 5 pages maximum) that includes an
introduction, results, a discussion section, references and at least 2 figures
3) A completed web template containing the main results from the research paper
4) An 8-minute presentation that each student will deliver to the class

Example project topics:
Can we design a new lens/transducer/antenna shape to improve classification of X?
What is the tradeoff between image resolution and classification accuracy for X?
Can we determine an optimal set of colors to improve fluorophore distinguishability?
If we capture 2 images that are overlapped on one sensor, what is the best way to pre-blur them to then be able to tell them
apart? Or to be able to classify them together?
If we just had a few sensors, how should be arrange them e.g. a mask to be able to predict the position of X?
Is there some optimal shift-variant blur that we can to use for a particular task?
Or, given a shift-variant PSF image, can we establish a good deconvolution using locally connected layers?
What is the optimal way to layout filters on a sensor to capture a color image for classification? Or an HDR image?
HDR image generation with filters over pixels – what is optimal design?
What if we could make a sensor with different sized pixels – how should they be laid out to achieve the best X?

Class project – what are the first steps?
1. Think about it!
2. Discuss with your friends/others in the class (feel free to use Slack!)
3. Schedule a short 15 meeting with me:
•

Friday 3/1, 3:30pm – 6:00pm

•

Next Monday 3/4, 1:00pm – 3:30pm

•

Next Wednesday 3/6, 10:00am – 1:00pm

4. Start to write-up a proposal
•

General aim: 1 paragraph with specification of physical layer

•

Discussion: (a) data source(s), (b) expected simulations, (c) expected CNN, (d)
quantitative analysis of physical layer (comparison, plot, etc).

•

Project proposal due date: Thursday March 7, 2019

•

Revised project proposal due date: Tuesday March 19, 2019

Example project topics:
Can we design a new lens/transducer/antenna shape to improve classification of X?
What is the tradeoff between image resolution and accuracy for X (classification, segmentation, etc.)? What if we had access to n
low-resolution cameras – how might we position them to get the best performance?
Can we determine an optimal set of colors to improve fluorophore distinguishability?
If we capture 2 images that are overlapped on one sensor, what is the best way to pre-blur them to then be able to tell them
apart? Or to be able to classify them together?
If we just had a few sensors, how should be arrange them e.g. a mask to be able to predict the position of X?
Is there some optimal shift-variant blur that we can to use for a particular task?
Or, given a shift-variant PSF image, can we establish a good deconvolution using locally connected layers?
What is the optimal way to layout filters on a sensor to capture a color image for classification? Or an HDR image?
HDR image generation with filters over pixels – what is optimal design?
What if we could make a sensor with different sized pixels – how should they be laid out to achieve the best X?

What is the tradeoff between image resolution and accuracy for image segmentation? What if we had
access to n low-resolution cameras – how might we position them to get the best performance?
I propose to simulate the classification performance of a new
type of microscope, which will have 3 different lenses and
sensors. Each lens and sensor will capture an image of a flat
object from a unique angular perspective, and the image
classification will be performed with all of the data. The physical
parameter that I will optimize is the angle of tilt of each lens
with respect to the object to maximize classification accuracy.

What is the tradeoff between image resolution and accuracy for image segmentation? What if we had
access to n low-resolution cameras – how might we position them to get the best performance?

Optimize: a1-3 to max. classification
accuracy of imaging system
Classification CNN

Image 1

Image 2

a1

a2

a3

Image 3

I propose to simulate the classification performance of a new
type of microscope, which will have 3 different lenses and
sensors. Each lens and sensor will capture an image of a flat
object from a unique angular perspective, and the image
classification will be performed with all of the data. The physical
parameter that I will optimize is the angle of tilt of each lens
with respect to the object to maximize classification accuracy.

What is the tradeoff between image resolution and accuracy for image segmentation? What if we had
access to n low-resolution cameras – how might we position them to get the best performance?

Optimize: a1-3 to max. classification
accuracy of imaging system
Classification CNN

Image 1

Image 2

a1

a2

a3

I propose to simulate the classification performance of a new
type of microscope, which will have 3 different lenses and
sensors. Each lens and sensor will capture an image of a flat
object from a unique angular perspective, and the image
classification will be performed with all of the data. The physical
parameter that I will optimize is the angle of tilt of each lens
with respect to the object to maximize classification accuracy.
Dataset: 12,500 images of 4 types of blood cell
https://www.kaggle.com/paultimothymooney/blood-cells

Image 3

Simulation: Treat each image as a thin 2D object and is coherently illuminated.
Assume each camera captures a unique component of the object spectrum, which
will vary as a function of a1-3. Start by neglecting size and shape of each camera.
CNN: Digital layer: Alexnet. Physical layer: simulate object spectrum, sample object
spectrum re-centered by angle a1-3 (which are weight variables), form images and
classify them together
Quantitative analysis: I will plot classification performance as a function of the
number of allowed cameras for a fixed CNN architecture, and will also compare the
classification performance to the case of a single image



Source Exif Data:
File Type                       : PDF
File Type Extension             : pdf
MIME Type                       : application/pdf
PDF Version                     : 1.6
Linearized                      : Yes
Create Date                     : 2019:02:26 14:05:58-05:00
Creator                         : PowerPoint
Modify Date                     : 2019:02:26 14:06:12-05:00
XMP Toolkit                     : Adobe XMP Core 5.4-c006 80.159825, 2016/09/16-03:31:08
Metadata Date                   : 2019:02:26 14:06:12-05:00
Creator Tool                    : PowerPoint
Format                          : application/pdf
Title                           : lecture_14_object_detection_segmentation
Document ID                     : uuid:db020450-7ee8-f944-b581-088ba1187d89
Instance ID                     : uuid:e23fcc96-96e6-2e4a-a945-bd128e2959c6
Producer                        : Mac OS X 10.12.4 Quartz PDFContext
Page Count                      : 8
EXIF Metadata provided by EXIF.tools

Navigation menu