Lecture_14_object_detection_segmentation Project Proposal Instructions
User Manual:
Open the PDF directly: View PDF .
Page Count: 8
Download | ![]() |
Open PDF In Browser | View PDF |
Class project details • Can work alone or in a group (up to 4 people), required effort will scale with # of people • Select a “base” dataset (online, or from a list I’ll make) • Simulate parameters of a physical (imaging) system with base dataset • Train deep neural net with simulated dataset • Report results Class project details • Can work alone or in a group (up to 4 people), required effort will scale with # of people • Select a “base” dataset (online, or from a list I’ll make) • Simulate parameters of a physical (imaging) system with base dataset • Train deep neural net with simulated dataset • Report results What you’ll need to submit: 1) The project’s source code 2) A short research-style paper (3 pages minimum, 5 pages maximum) that includes an introduction, results, a discussion section, references and at least 2 figures 3) A completed web template containing the main results from the research paper 4) An 8-minute presentation that each student will deliver to the class Example project topics: Can we design a new lens/transducer/antenna shape to improve classification of X? What is the tradeoff between image resolution and classification accuracy for X? Can we determine an optimal set of colors to improve fluorophore distinguishability? If we capture 2 images that are overlapped on one sensor, what is the best way to pre-blur them to then be able to tell them apart? Or to be able to classify them together? If we just had a few sensors, how should be arrange them e.g. a mask to be able to predict the position of X? Is there some optimal shift-variant blur that we can to use for a particular task? Or, given a shift-variant PSF image, can we establish a good deconvolution using locally connected layers? What is the optimal way to layout filters on a sensor to capture a color image for classification? Or an HDR image? HDR image generation with filters over pixels – what is optimal design? What if we could make a sensor with different sized pixels – how should they be laid out to achieve the best X? Class project – what are the first steps? 1. Think about it! 2. Discuss with your friends/others in the class (feel free to use Slack!) 3. Schedule a short 15 meeting with me: • Friday 3/1, 3:30pm – 6:00pm • Next Monday 3/4, 1:00pm – 3:30pm • Next Wednesday 3/6, 10:00am – 1:00pm 4. Start to write-up a proposal • General aim: 1 paragraph with specification of physical layer • Discussion: (a) data source(s), (b) expected simulations, (c) expected CNN, (d) quantitative analysis of physical layer (comparison, plot, etc). • Project proposal due date: Thursday March 7, 2019 • Revised project proposal due date: Tuesday March 19, 2019 Example project topics: Can we design a new lens/transducer/antenna shape to improve classification of X? What is the tradeoff between image resolution and accuracy for X (classification, segmentation, etc.)? What if we had access to n low-resolution cameras – how might we position them to get the best performance? Can we determine an optimal set of colors to improve fluorophore distinguishability? If we capture 2 images that are overlapped on one sensor, what is the best way to pre-blur them to then be able to tell them apart? Or to be able to classify them together? If we just had a few sensors, how should be arrange them e.g. a mask to be able to predict the position of X? Is there some optimal shift-variant blur that we can to use for a particular task? Or, given a shift-variant PSF image, can we establish a good deconvolution using locally connected layers? What is the optimal way to layout filters on a sensor to capture a color image for classification? Or an HDR image? HDR image generation with filters over pixels – what is optimal design? What if we could make a sensor with different sized pixels – how should they be laid out to achieve the best X? What is the tradeoff between image resolution and accuracy for image segmentation? What if we had access to n low-resolution cameras – how might we position them to get the best performance? I propose to simulate the classification performance of a new type of microscope, which will have 3 different lenses and sensors. Each lens and sensor will capture an image of a flat object from a unique angular perspective, and the image classification will be performed with all of the data. The physical parameter that I will optimize is the angle of tilt of each lens with respect to the object to maximize classification accuracy. What is the tradeoff between image resolution and accuracy for image segmentation? What if we had access to n low-resolution cameras – how might we position them to get the best performance? Optimize: a1-3 to max. classification accuracy of imaging system Classification CNN Image 1 Image 2 a1 a2 a3 Image 3 I propose to simulate the classification performance of a new type of microscope, which will have 3 different lenses and sensors. Each lens and sensor will capture an image of a flat object from a unique angular perspective, and the image classification will be performed with all of the data. The physical parameter that I will optimize is the angle of tilt of each lens with respect to the object to maximize classification accuracy. What is the tradeoff between image resolution and accuracy for image segmentation? What if we had access to n low-resolution cameras – how might we position them to get the best performance? Optimize: a1-3 to max. classification accuracy of imaging system Classification CNN Image 1 Image 2 a1 a2 a3 I propose to simulate the classification performance of a new type of microscope, which will have 3 different lenses and sensors. Each lens and sensor will capture an image of a flat object from a unique angular perspective, and the image classification will be performed with all of the data. The physical parameter that I will optimize is the angle of tilt of each lens with respect to the object to maximize classification accuracy. Dataset: 12,500 images of 4 types of blood cell https://www.kaggle.com/paultimothymooney/blood-cells Image 3 Simulation: Treat each image as a thin 2D object and is coherently illuminated. Assume each camera captures a unique component of the object spectrum, which will vary as a function of a1-3. Start by neglecting size and shape of each camera. CNN: Digital layer: Alexnet. Physical layer: simulate object spectrum, sample object spectrum re-centered by angle a1-3 (which are weight variables), form images and classify them together Quantitative analysis: I will plot classification performance as a function of the number of allowed cameras for a fixed CNN architecture, and will also compare the classification performance to the case of a single image
Source Exif Data:
File Type : PDF File Type Extension : pdf MIME Type : application/pdf PDF Version : 1.6 Linearized : Yes Create Date : 2019:02:26 14:05:58-05:00 Creator : PowerPoint Modify Date : 2019:02:26 14:06:12-05:00 XMP Toolkit : Adobe XMP Core 5.4-c006 80.159825, 2016/09/16-03:31:08 Metadata Date : 2019:02:26 14:06:12-05:00 Creator Tool : PowerPoint Format : application/pdf Title : lecture_14_object_detection_segmentation Document ID : uuid:db020450-7ee8-f944-b581-088ba1187d89 Instance ID : uuid:e23fcc96-96e6-2e4a-a945-bd128e2959c6 Producer : Mac OS X 10.12.4 Quartz PDFContext Page Count : 8EXIF Metadata provided by EXIF.tools