合肥生活安徽新聞合肥交通合肥房產生活服務合肥教育合肥招聘合肥旅游文化藝術合肥美食合肥地圖合肥社保合肥醫院企業服務合肥法律

        CS 551代寫、c/c++設計編程代做
        CS 551代寫、c/c++設計編程代做

        時間:2024-11-28  來源:合肥網hfw.cc  作者:hfw.cc 我要糾錯



        CS 551 Systems Programming, Fall 2024
        Programming Project 2
        In this project we are going to simulate the MapReduce framework on a single machine using
        multi-process programming.
        1 Introduction
        In 2004, Google (the paper “MapReduce: Simplified Data Processing on Large Clusters” by J.
        Dean and S. Ghemawat) introduced a general programming model for processing and generating
        large data sets on a cluster of computers.
        The general idea of the MapReduce model is to partition large data sets into multiple splits,
        each of which is small enough to be processed on a single machine, called a worker. The data
        splits will be processed in two phases: the map phase and the reduce phase. In the map phase, a
        worker runs user-defined map functions to parse the input data (i.e., a split of data) into multiple
        intermediate key/value pairs, which are saved into intermediate files. In the reduce phase, a
        (reduce) worker runs reduce functions that are also provided by the user to merge the intermediate
        files, and outputs the result to result file(s).
        We now use a small data set (the first few lines of a famous poem by Robert Frost, see Figure
        1) to explain to what MapReduce does.
        Figure 1: A small data set to be processed by MapReduce.
        To run MapReduce, we first split the dataset into small pieces. For this example, we will split
        the dataset by the four lines of the poem (Figure 2).
        Figure 2: Partitioning the input data set into multiple splits.
        The MapReduce framework will have four workers (in our project, the four workers are four
        processes that are forked by the main program. In reality, they will be four independent machines)
        to work on the four splits (each worker is working on a split). These four map worker will each
        run a user-defined map function to process the split. The map function will map the input into
        a series of (key, value) pairs. For this example, let the map function simply count the number of
        each letter (A-Z) in the data set.
        Figure 3: The outputs of the map phase, which are also the inputs to the reduce phase.
        The map outputs in our example are shown in Figure 3. They are also the inputs for the
        reduce phase. In the reduce phase, a reduce worker runs a user-defined reduce function to merge
        the intermediate results output by the map workers, and generates the final results (Figure 4).
        Figure 4: The final result
        2 Simulating the MapReduce with multi-process programming
        2.1 The base code
        Download the base code from the Brightspace. You will need to add your implementation into
        this base code. The base code also contains three input data sets as examples.
        2.2 The working scenario
        In this project, we will use the MapReduce model to process large text files. The input will be a
        file that contains many lines of text. The base code folder contains three example input data files.
        We will be testing using the example input data files, or data files in similar format.
        A driver program is used to accept user inputs and drive the MapReduce processing. The
        main part of driver program is already implemented in main.c. You will need to complete the
        mapreduce() function, which is defined in mapreduce.c and is called by the driver program.
        A Makefile has already been given. Running the Makefile can give you the executable of the driver
        program, which is named as “run-mapreduce”. The driver program is used in the following way:
        ./run-mapreduce "counter"|"finder" file_path split_num [word_to_find]
        where the arguments are explained as follows.
        • The first argument specifies the type of the task, it can be either the “Letter counter” or
        the “Word conter” (explained later).
        • The second argument “file path” is the path to the input data file.
        • The third argument “split num” specifies how many splits the input data file should be
        partitioned into for the map phase.
        • The fourth argument is used only for the “Word finder” task. This argument specifies the
        word that the user is trying to find in the input file.
        The mapreduce() function will first partition the input file into N roughly equal-sized splits,
        where N is determined by the split num argument of the driver program. Note that the sizes of
        each splits do not need to be exactly the same, otherwise a word may be divided into two different
        splits.
        Then the mapreduce() forks one worker process per data split, and the worker process will
        run the user-defined map function on the data split. After all the splits have been processed, the
        first worker process forked will also need to run the user-defined reduce function to process all the
        intermediate files output by the map phase. Figure 5 below gives an example about this process.
        split 0
        split 1
        split 2
        Driver
        Program
        map
        worker 0
        reduce
        worker
        map
        worker 2
        map
        worker 3
        “mr-0.itm”
        “mr-1.itm”
        “mr-2.itm”
        “mr-3.itm”
        map
        worker 1
        (1) partition
        (2) fork
        (3) userdefined
        map
        (5) userdefined
        reduce
        “mr.rst”
        Input
        data file
        Intermediate
        files
        Result
        file
        PID=1001
        PID=1002
        PID=1003
        PID=1004
        PID=1001
        split 3
        Figure 5: An example of the working scenario.
        2.3 The two tasks
        The two tasks that can be performed by the driver program are described as follows.
        The “Letter counter” task is similar to the example we showed in Section 1, which is counting
        the number of occurrence of the 26 letters in the input file. The difference is the intermediate file
        and the final result file should be written in the following format:
        A number-of-occurrences
        B number-of-occurrences
        ...
        Y number-of-occurrences
        Z number-of-occurrences
        The “Word finder” task is to find the word provided by user (specified by the “word to find”
        argument of the driver program) in the input file, and outputs to the result file all the lines that
        contain the target word in the same order as they appear in the input file. For this task, you
        should implement the word finder as a whole word match, meaning that the function should only
        recognize complete words that match exactly(case-sensitive) with the specified search terms. And
        if multiple specified words are found in the same line, you only need to output that line once.
        2.4 Other requirements
        • Besides the mapreduce() function defined in mapreduce.c, you will also need to complete the map/reduce functions of the two tasks (in usr functions.c.)
        • About the interfaces listed in “user functions.h” and “mapreduce.h”:
        – Do not change any function interfaces.
        – Do not change or delete any fields in the structure interfaces (but you may add additional fields in the structure interface if necessary).
        The above requirements allow the TA to test your implementations of worker logic and user
        map/reduce functions separately. Note that violation to these requirements will result in 0
        point for this project.
        • Use fork() to spawn processes.
        • Be careful to avoid fork bomb (check on Wikipedia if you are not familiar with it). A fork
        bomb will result in 0 point for this project.
        • The fd in the DATA SPLIT structure should be a file descriptor to the original input data
        file.
        • The intermediate file output by the first map worker process should be named as “mr-0.itm”,
        the intermediate file by the second map worker process should be named as “mr-1.itm”, ...
        The result file is named as “mr.rst” (already done in main.c).
        • Program should not automatically delete the intermediate files once they are created. They
        will be checked when grading. But your submission should not contain any intermediate
        files as they should be created dynamically.
        3 Submit your work
        Compress the files: compress your README file, all the files in the base code folder, and
        any additional files you add into a ZIP file. Name the ZIP file based on your BU email ID. For
        example, if your BU email is “abc@binghamton.edu”, then the zip file should be “proj2 abc.zip”.
        Submission: submit the ZIP file to Brightspace before the deadline.
        3.1 Grading guidelines
        (1) Prepare the ZIP file on a Linux machine. If your zip file cannot be uncompressed, 5 points
        off.
        (2) If the submitted ZIP file/source code files included in the ZIP file are not named as specified
        above (so that it causes problems for TA’s automated grading scripts), 10 points off.
        (3) If the submitted code does not compile:
        1 TA will try to fix the problem (for no more than 3 minutes);
        2 if (problem solved)
        3 1%-10% points off (based on how complex the fix is, TA’s discretion);
        4 else
        5 TA may contact the student by email or schedule a demo to fix the problem;
        6 if (problem solved)
        7 11%-20% points off (based on how complex the fix is, TA’s discretion);
        8 else
        9 All points off;
        So in the case that TA contacts you to fix a problem, please respond to TA’s email promptly
        or show up at the demo appointment on time; otherwise the line 9 above will be effective.
        (4) If the code is not working as required in this spec, the TA should take points based on the
        assigned full points of the task and the actual problem.
        (5) Lastly but not the least, stick to the collaboration policy stated in the syllabus: you may
        discuss with your fellow students, but code should absolutely be kept private.

        請加QQ:99515681  郵箱:99515681@qq.com   WX:codinghelp




         

        掃一掃在手機打開當前頁
      1. 上一篇:COMP4134代做、Java程序語言代寫
      2. 下一篇:中國最低調的隱形富豪起名大師顏廷利:全球點贊之父,國際享有盛譽
      3. 無相關信息
        合肥生活資訊

        合肥圖文信息
        挖掘機濾芯提升發動機性能
        挖掘機濾芯提升發動機性能
        戴納斯帝壁掛爐全國售后服務電話24小時官網400(全國服務熱線)
        戴納斯帝壁掛爐全國售后服務電話24小時官網
        菲斯曼壁掛爐全國統一400售后維修服務電話24小時服務熱線
        菲斯曼壁掛爐全國統一400售后維修服務電話2
        美的熱水器售后服務技術咨詢電話全國24小時客服熱線
        美的熱水器售后服務技術咨詢電話全國24小時
        海信羅馬假日洗衣機亮相AWE  復古美學與現代科技完美結合
        海信羅馬假日洗衣機亮相AWE 復古美學與現代
        合肥機場巴士4號線
        合肥機場巴士4號線
        合肥機場巴士3號線
        合肥機場巴士3號線
        合肥機場巴士2號線
        合肥機場巴士2號線
      4. 幣安app官網下載 短信驗證碼 丁香花影院

        關于我們 | 打賞支持 | 廣告服務 | 聯系我們 | 網站地圖 | 免責聲明 | 幫助中心 | 友情鏈接 |

        Copyright © 2024 hfw.cc Inc. All Rights Reserved. 合肥網 版權所有
        ICP備06013414號-3 公安備 42010502001045

        主站蜘蛛池模板: 人妻无码一区二区不卡无码av| 国产av天堂一区二区三区| 亚洲爽爽一区二区三区| 亚洲国产激情一区二区三区| 中文字幕乱码一区二区免费| 午夜性色一区二区三区不卡视频 | 亚洲综合在线成人一区| 午夜无码视频一区二区三区| 国产乱人伦精品一区二区在线观看| 亚洲国产精品一区二区第一页免 | 免费精品一区二区三区在线观看 | 久99精品视频在线观看婷亚洲片国产一区一级在线 | 久久久av波多野一区二区| 中文字幕在线无码一区| 一区二区高清在线观看| 无码精品久久一区二区三区| 无码日韩AV一区二区三区| 午夜视频在线观看一区二区| 曰韩人妻无码一区二区三区综合部 | 亚洲一区二区三区四区在线观看| 一区二区三区中文| 一区二区三区福利视频| 丰满岳妇乱一区二区三区| 视频在线观看一区二区三区| 国产一区韩国女主播| 能在线观看的一区二区三区| 一区二区高清在线| 中文字幕永久一区二区三区在线观看| 日本免费精品一区二区三区| 一本色道久久综合一区| 中文字幕Av一区乱码| 亚洲国产成人一区二区精品区| 久久99精品波多结衣一区| 蜜桃臀无码内射一区二区三区| 精品无码一区在线观看| 亚洲AV无码国产一区二区三区 | 精品在线一区二区| 亚洲色欲一区二区三区在线观看| 无码人妻精品一区二区三区99仓本 | 中文字幕在线一区| 亚洲AV无码一区二区乱子伦|