基于GPU的網(wǎng)絡(luò)流量特征提取并行算法設(shè)計(jì)與實(shí)現(xiàn)
發(fā)布時(shí)間:2019-02-14 07:04
【摘要】:網(wǎng)絡(luò)流量分類技術(shù)在增強(qiáng)網(wǎng)絡(luò)可控性以及加強(qiáng)網(wǎng)絡(luò)管理方面都發(fā)揮著重要的作用。隨著網(wǎng)絡(luò)應(yīng)用的層出不窮,對(duì)實(shí)時(shí)、準(zhǔn)確的流量分類技術(shù)提出了更高的要求,使得近年來(lái)研究者大量引入機(jī)器學(xué)習(xí)領(lǐng)域知識(shí)來(lái)處理流量分類問(wèn)題,取得了較好的分類效果。但是,特征提取作為機(jī)器學(xué)習(xí)分類算法中一個(gè)重要環(huán)節(jié),在處理大數(shù)據(jù)流量時(shí)因其計(jì)算復(fù)雜度較高、耗時(shí)過(guò)長(zhǎng)已成為制約機(jī)器學(xué)習(xí)算法應(yīng)用于實(shí)時(shí)流量分類的主要瓶頸。 近年來(lái),GPU硬件體系結(jié)構(gòu)的快速發(fā)展使其浮點(diǎn)運(yùn)算和并行計(jì)算能力遠(yuǎn)遠(yuǎn)超過(guò)了CPU,在大規(guī)模并行處理和科學(xué)計(jì)算等方面取得了廣泛應(yīng)用。特別是NVIDIA公司CUDA編程模型的推出,提供了豐富的API函數(shù),使其能夠更好地發(fā)揮GPU并行處理能力。 本文首先介紹了GPU在體系結(jié)構(gòu)和編程方式上與傳統(tǒng)CPU的不同。其次,對(duì)串行特征提取算法執(zhí)行流程進(jìn)行了介紹,并從串行算法每部分的計(jì)算任務(wù)大小和特點(diǎn)入手對(duì)算法的可并行性進(jìn)行了逐一地分析。在此基礎(chǔ)上,采用CUDA編程語(yǔ)言設(shè)計(jì)實(shí)現(xiàn)了并行特征提取算法,并利用流化技術(shù)和GPU異構(gòu)執(zhí)行的特點(diǎn)對(duì)并行算法進(jìn)行了優(yōu)化。最后,通過(guò)實(shí)驗(yàn)對(duì)本文提出的并行算法及其優(yōu)化方案在LINUX平臺(tái)下進(jìn)行了測(cè)試和驗(yàn)證。實(shí)驗(yàn)結(jié)果表明,在進(jìn)行大數(shù)據(jù)量的網(wǎng)絡(luò)流特征提取時(shí),優(yōu)化后的并行算法相比于串行算法可以達(dá)到2倍以上的加速比,取得了顯著的性能優(yōu)勢(shì)。
[Abstract]:Network traffic classification plays an important role in enhancing network controllability and network management. With the continuous emergence of network applications, higher requirements for real-time and accurate traffic classification techniques have been put forward. In recent years, researchers have introduced a large number of machine learning domain knowledge to deal with traffic classification problems, and achieved better classification results. However, as an important part of machine learning classification algorithm, feature extraction has become the main bottleneck in the application of machine learning algorithm in real-time traffic classification because of its high computational complexity and long time consuming. In recent years, with the rapid development of GPU hardware architecture, its floating-point computing and parallel computing capabilities have far exceeded the extensive application of CPU, in large-scale parallel processing and scientific computing. Especially, the introduction of CUDA programming model of NVIDIA Company provides abundant API functions, which makes it better exert the ability of GPU parallel processing. This paper first introduces the differences between GPU and traditional CPU in architecture and programming. Secondly, the execution flow of the serial feature extraction algorithm is introduced, and the parallelism of the algorithm is analyzed one by one from the calculation task size and characteristics of each part of the serial algorithm. On this basis, the parallel feature extraction algorithm is designed and implemented by using CUDA programming language, and the parallel algorithm is optimized by using the fluidization technology and the characteristics of GPU heterogeneous execution. Finally, the proposed parallel algorithm and its optimization scheme are tested and verified on LINUX platform through experiments. The experimental results show that the optimized parallel algorithm can achieve a speedup ratio of more than 2 times compared with the serial algorithm, and obtain a significant performance advantage.
【學(xué)位授予單位】:內(nèi)蒙古大學(xué)
【學(xué)位級(jí)別】:碩士
【學(xué)位授予年份】:2014
【分類號(hào)】:TP181;TP393.06
本文編號(hào):2421942
[Abstract]:Network traffic classification plays an important role in enhancing network controllability and network management. With the continuous emergence of network applications, higher requirements for real-time and accurate traffic classification techniques have been put forward. In recent years, researchers have introduced a large number of machine learning domain knowledge to deal with traffic classification problems, and achieved better classification results. However, as an important part of machine learning classification algorithm, feature extraction has become the main bottleneck in the application of machine learning algorithm in real-time traffic classification because of its high computational complexity and long time consuming. In recent years, with the rapid development of GPU hardware architecture, its floating-point computing and parallel computing capabilities have far exceeded the extensive application of CPU, in large-scale parallel processing and scientific computing. Especially, the introduction of CUDA programming model of NVIDIA Company provides abundant API functions, which makes it better exert the ability of GPU parallel processing. This paper first introduces the differences between GPU and traditional CPU in architecture and programming. Secondly, the execution flow of the serial feature extraction algorithm is introduced, and the parallelism of the algorithm is analyzed one by one from the calculation task size and characteristics of each part of the serial algorithm. On this basis, the parallel feature extraction algorithm is designed and implemented by using CUDA programming language, and the parallel algorithm is optimized by using the fluidization technology and the characteristics of GPU heterogeneous execution. Finally, the proposed parallel algorithm and its optimization scheme are tested and verified on LINUX platform through experiments. The experimental results show that the optimized parallel algorithm can achieve a speedup ratio of more than 2 times compared with the serial algorithm, and obtain a significant performance advantage.
【學(xué)位授予單位】:內(nèi)蒙古大學(xué)
【學(xué)位級(jí)別】:碩士
【學(xué)位授予年份】:2014
【分類號(hào)】:TP181;TP393.06
【參考文獻(xiàn)】
相關(guān)期刊論文 前3條
1 徐鵬;劉瓊;林森;;改進(jìn)的對(duì)等網(wǎng)絡(luò)流量傳輸層識(shí)別方法[J];計(jì)算機(jī)研究與發(fā)展;2008年05期
2 徐鵬;劉瓊;林森;;基于支持向量機(jī)的Internet流量分類研究[J];計(jì)算機(jī)研究與發(fā)展;2009年03期
3 熊剛;孟姣;曹自剛;王勇;郭莉;方濱興;;網(wǎng)絡(luò)流量分類研究進(jìn)展與展望[J];集成技術(shù);2012年01期
,本文編號(hào):2421942
本文鏈接:http://www.sikaile.net/guanlilunwen/ydhl/2421942.html
最近更新
教材專著