合肥生活安徽新聞合肥交通合肥房產(chǎn)生活服務(wù)合肥教育合肥招聘合肥旅游文化藝術(shù)合肥美食合肥地圖合肥社保合肥醫(yī)院企業(yè)服務(wù)合肥法律

        CS 435代做、代寫Matlab編程設(shè)計(jì)

        時(shí)間:2024-03-05  來源:合肥網(wǎng)hfw.cc  作者:hfw.cc 我要糾錯(cuò)



        CS 435 - Computational Photography
        Final Project - Panoramic Stitching
        YOU MAY WORK WITH A PARTNER IF YOU LIKE!!!
        But if you do so, look at the additional information you need to provide in your submission
        (stated at the end of the document).
        Introduction
        For our final assignment, we’ll attack the problem of creating a panoramic photo. This will require
        several ideas from this course, including:
         Least Squares Estimate (LSE) for Transformation Matrix Discovery
         Projection
         Blending
         Interest Point Discovery (subsampling, gradients, edges)
         Respresentation (feature extraction)
         Feature Matching (point correspondences).
        You may use functions that you were allowed to use in prior assignments. In particular things like
        edge, imgradientxy, imgausfilt, etc.. However you may not use Matlab functions to do the new things
        in this assignment. In particular, functions that might find keypoints and/or do transformations
        (like imtransform, imregionalmask, imwarp, etc.. In additino, you cannot use anything from the
        Computer Vision or Machine Learning toolboxes. This is not an exhaustive list, but hopefully you
        get the idea. If in doubt, ask your instructor!
        The Dataset
        For the programming component of this assignment, take two pictures, one slightly offset from the
        other (via rotation and/or translation). Make sure that the two images have significant overlap of
        content.
        1
        Grading
        Hard Coded Correspondences 10pts
        Panoramic using hard-coded correspondences 30pts
        Image Pyramids 10pts
        Extrema Points 10pts
        Keypoint Matching 10pts
        Automatic Stitching 10pts
        Success on Additional Tests 12pts
        Report quality an ease of running code 8pts
        TOTAL 100pts
        Table 1: Grading Rubric
        2
        1 (10 points) Hard Coding Point Correspondences
        Let’s start off by hard coding some point correspondences. Look at each image and choose four
        point correspondences. Do not make this process interactive. Hard code the coordinates at the top
        of your script.
        Display the images side-by-side (as one image) with the point correspondences color coded as dots
        in the image. An example can be found in Figure 1.
        Figure 1: Manual Correspondences
        3
        2 (30 points) Compute Transformation Matrix, Project, and
        Blend!
        Next, use the four points you identified in the previous part to compute the transformation matrix
        that maps one image to the other. You can determine which image you want to be the “base” image.
        After determining the transformation matrix, we need to determine the dimensions of the new combined image. The height of this image should be the maximum of the base image’s height or the
        maximum projected y value from the other image. The width will be equal to the maximum of the
        base image’s width or the maximum projected x value from the other image.
        Finally we need to populate our new image with pixel(s) from the base and projected images. To do
        this, go through each location in your new image and grab the corresponding pixels from the base
        and/or projected image (you’ll need to determine where, if anywhere, these come from). If both
        images map to that location, you’ll want to blend them (using a technique of your choosing).
        An example can be found in Figure 2.
        Figure 2: Stitched images using manual correspondences
        4
        3 (10 points) Create Scale-Space Image Pyramids
        Now on to the tough(er) stuff! We want to automate all this!
        The first step is to automatically identify locations of interest. To do this we’ll find the stable local
        maximas in scale-space for each image. And the first step of that is to create image pyramids!
        Here are some hyperparameters we’ll use to create our image pyramids:
        ˆ Find the extremas in grayscale.
        ˆ Create five scales per octave.
        ˆ The initial scale will have a standard deviation of σ0 = 1.6.
        ˆ Each subsequent scale will have a σ value that is k =

        2 times larger than the previous.
        ˆ Each Gaussian kernel will have a width and height that is three times the filter’s σ value, i.e.
        w = ⌈3σ⌉.
        ˆ Create four octaves, each 1/4 of the size of the previous octave, obtained by subsampling ever
        other row and column of the previous column (no interpolation).
        In general, given octave n and scale m, you can compute σ as:
        σ = 2n−1
        k
        m−1σ0
        In your report show all the images for each octave for one of you images. Something similar to Figure
        3.
        5
        Figure 3: Image Pyramid
        6
        4 (10 points) Finding the Local Maximas
        Next, for each octave of each image, locate the local maxima, as discussed in class. These locations
        then need to be in terms of the original image’s size (i.e. the first octave), which can be done by
        multiplying their locations by 2n−1
        , where again n is the current octave.
        After identifying all the extrams, we want to remove the unstable ones, i.e. those that are edge pixels
        and/or in areas of low contrast. To do this:
        ˆ Find edge pixels use Matlab’s edge function. This will return a binary image (where a value of
        one indicates that the pixel is an edge pixel). Use that (perhaps along with Matlab’s find and
        setdiff functions) to eliminate extremas that are also edge pixels.
        ˆ We will also eliminate extremas that are too close to the border of the image. You can determine
        what “too close” means, but your choice will likely be related to your descriptor decision in
        Part 5 (and how large of a region around they keypoints you’ll use to form the descriptors).
        ˆ Finally, for each remaining extrema, compute the standard deviation of a patch around it. If
        this standard deviation is less than some threshold, then the patch has low contrast and thus
        should be eliminated from the extrema list. Once again, you can decide on the size of the patch
        and the threshold based on experimentation.
        For your report, provide two images for each input image. One with all the extremas superimposed
        on it (indicated by red circles), and one after unstable extremas were removed. As an example, see
        Figures 4-5.
        Figure 4: All extrema points
        7
        Figure 5: Pruned extrema points
        5 (10 points) Keypoint Description and Matching
        For each remaining extrema/keypoint in each image, we’ll want to extract a descriptor and then
        match the descriptors from one image to ones in the other. To compare keypoints, you will have to
        determine what distance or similarity measurement to use. Common distance ones are Eucliden and
        Manhattan. Common similarity ones are Cosine, Gaussian, and Histogram Intersection.
        The following sections discuss strategies for describing keypoint regions (descriptor extraction) and
        keypoint matching.
        5.1 Descriptors
        Given the constraints/assumptions of the problem, describing a patch around a keypoint using the
        RGB values will likely work well (since it encodes both color and positional information). Thus,
        if we had 9 × 9 region around a keypoint, we could describe that keypoint with a vector of size
        9 × 9 × 3 = 243 values. However, feel free to experiment with other descriptors (SIFTs, Local
        Histograms, Local GISTs, etc..).
        5.2 Keypoint Correspondences
        To find keypoint correspondences between images, we’ll make a few problem-specific assumptions:
        ˆ Correspondences should have roughly the same y value.
        ˆ The camera was rotated and/or translated right to obtain the second image.
        Our general keypoint matching strategy will be:
        1. For each keypoint in the first image, find the best match (using the distance or similarity
        measurement of your choice) in the second image that satisfies the aforementioned constraints.
        Call this set C1.
        2. For each keypoint in the second image, find the best match (using the distance or similarity
        measurement of your choice) in the first image that satisfies the aforementioned constraints.
        Call this set C2.
        3. Computer the set intersection of these two sets: C = C1 ∩ C2.
        8
        4. Remove from C all correspondences that have a distance above some threshold (or if you use
        similarity, below some threshold).
        For visualization (and your report), draw lines between a few matching keypoints, as seen in Figure
        6.
        Figure 6: Some Point Correspondences
        9
        6 (10 points) Find the Transformation Matrix via RANSAC
        and Stitch
        Finally we want to use the keypoint correspondences to compute a transformation matrix that we
        can then use to auto-stitch our images.
        However, as you may have noticed, many of the point correspondences might not be correct :(. So
        instead we’ll use a RANSAC RANdom SAmpling Consensus strategy.
        To perform RANSAC for our panoramic stitching:
        1. For experiments 1 through N (you choose N)
        (a) Select four correspondences at random.
        (b) Compute the transformation matrix using these correspondences.
        (c) Using the discovered transformation matrix, count how many point correspondences (among
        all of them) would end up within a few pixels of one another after projection.
        2. Keep the transformation matrix the resulting in the largest number of point correspondences
        (among all of them) that ended up within a few pixels of one another after projection.
        Now use this transformation matrix to stitch your images!
        In your report:
        ˆ Draw lines between the keypoint coorespondences used to computer your final transformation
        matrix. See in Figure 7.
        ˆ Your final stitched image.
        10
        Figure 7: Point Correspondences for final transformation matrix
        7 (12 points) Additional Tests
        For the remaining points we’ll test your code against three other picture pairs. You will get 0-4
        points for each, depending on how well they stitched together.
        11
        Submission
        NOTE: that 8 points of your grade is based on being able to run your code easily.
        IN ADDITION: With your your submission, if you worked with someone else, let me know how
        evenly the work was split. If each contributed evenly it would be 50/50. I will use this information
        to adjust grades for pairs where one partner did more of the work.
        For your submission, upload to Blackboard a single zip file containing:
        1. PDF writeup that includes:
        (a) Visualization for Part 1
        (b) Stitched image for Part 2
        (c) Visualization for Part 3
        (d) Visualization for Part 4
        (e) Visualization for Part 5
        (f) Visualization and stitched image for Part 6
        2. A README text file (not Word or PDF) that explains
        ˆ Features of your program
        ˆ Name of your entry-point script
        ˆ Any useful instructions to run your script.
        3. Your source files
        請加QQ:99515681  郵箱:99515681@qq.com   WX:codehelp 

        掃一掃在手機(jī)打開當(dāng)前頁
      1. 上一篇:代寫CMPSC 221 UML and Class Creation
      2. 下一篇:COMP639代做、代寫Python/Java編程
      3. 無相關(guān)信息
        合肥生活資訊

        合肥圖文信息
        急尋熱仿真分析?代做熱仿真服務(wù)+熱設(shè)計(jì)優(yōu)化
        急尋熱仿真分析?代做熱仿真服務(wù)+熱設(shè)計(jì)優(yōu)化
        出評 開團(tuán)工具
        出評 開團(tuán)工具
        挖掘機(jī)濾芯提升發(fā)動(dòng)機(jī)性能
        挖掘機(jī)濾芯提升發(fā)動(dòng)機(jī)性能
        海信羅馬假日洗衣機(jī)亮相AWE  復(fù)古美學(xué)與現(xiàn)代科技完美結(jié)合
        海信羅馬假日洗衣機(jī)亮相AWE 復(fù)古美學(xué)與現(xiàn)代
        合肥機(jī)場巴士4號線
        合肥機(jī)場巴士4號線
        合肥機(jī)場巴士3號線
        合肥機(jī)場巴士3號線
        合肥機(jī)場巴士2號線
        合肥機(jī)場巴士2號線
        合肥機(jī)場巴士1號線
        合肥機(jī)場巴士1號線
      4. 短信驗(yàn)證碼 酒店vi設(shè)計(jì) NBA直播 幣安下載

        關(guān)于我們 | 打賞支持 | 廣告服務(wù) | 聯(lián)系我們 | 網(wǎng)站地圖 | 免責(zé)聲明 | 幫助中心 | 友情鏈接 |

        Copyright © 2025 hfw.cc Inc. All Rights Reserved. 合肥網(wǎng) 版權(quán)所有
        ICP備06013414號-3 公安備 42010502001045

        久久这里只有精品66| 99精品视频在线| 国产精品哟哟视频| 色婷婷精品大在线视频| 久久国语露脸国产精品电影| 亚洲AV成人精品网站在线播放| 国产美女亚洲精品久久久综合| 国产三级精品三级在线观看专1| 日本五区在线不卡精品| 日韩中文字幕在线| 日韩精品亚洲aⅴ在线影院| 国产精品99久久久| 欧洲精品免费一区二区三区| 久久精品国产亚洲av瑜伽| 亚洲国产精品自在自线观看| 国产成人精品1024在线| 国产91久久精品一区二区| 91久久精品国产成人久久| 久久亚洲精品无码aⅴ大香| 99re视频精品全部免费| 久久精品国产亚洲AV高清热| 99久久精品全部| 久久国产亚洲精品无码| 91久久精品国产91久久性色tv | 精品久久无码中文字幕| 亚洲精品福利网泷泽萝拉| 竹菊影视国产精品| 亚洲国产精品综合久久2007| 91久久精品国产91久久性色tv| 亚洲熟妇无码久久精品| 亚洲AV无码国产精品色| 国产人成精品香港三级在| 久久乐国产综合亚洲精品| 无码国产精品一区二区免费| 国产精品久久久久久久久鸭 | 青娱乐精品视频在线观看| 日韩久久精品一区二区三区| 中日韩精品视频在线观看| 国产精品青草久久久久福利99| 久久国产美女免费观看精品| 手机看片在线精品观看|