spark項目實踐

實踐目的

通過操作一個開源例子,學習大數據的架構 及基本的使用,各種概念。不涉及自編碼與創新。

環境搭建

需要建立 hadoop,hbase ,spark 等大數據環境

在10.30.2.5上建立六個docker , 分別對應 s141~s146 分別用於裝大數據環境,具體操作步驟 參考本人

hadoop-spark

https://blog.csdn.net/dualvencsdn/article/details/112007643?spm=1001.2014.3001.5501

habase

https://blog.csdn.net/dualvencsdn/article/details/112905925?spm=1001.2014.3001.5501

學會操作hbase

https://blog.csdn.net/dualvencsdn/article/details/113309385?spm=1001.2014.3001.5501
 

flume初步學習與使用

https://blog.csdn.net/qq_1018944104/article/details/85462011

/usr/local/flume/do.sh

kafka與zookeeper的使用與編程 

https://blog.csdn.net/dualvencsdn/article/details/105557575?ops_request_misc=%257B%2522request%255Fid%2522%253A%2522161234227816780269887393%2522%252C%2522scm%2522%253A%252220140713.130102334.pc%255Fblog.%2522%257D&request_id=161234227816780269887393&biz_id=0&utm_medium=distribute.pc_search_result.none-task-blog-2~blog~first_rank_v1~rank_blog_v1-1-105557575.pc_v1_rank_blog_v1&utm_term=kafka&spm=1018.2226.3001.4450

成果展現

 

 

操作記錄

/home/dualven/docker/*.jar  

start.sh -->start dockers

appendHost.sh-> add host ip for six hosts

 seeMessage.sh  ->see the message consumed by kafka

docker exec -it centos1122 bash

cd /usr/local/

see readme.txt

 代碼

https://codechina.csdn.net/dualvenorg/sparkstreaming.git

參考資料

https://blog.csdn.net/qq_41955099/article/details/88959996?utm_medium=distribute.pc_relevant.none-task-blog-BlogCommendFromMachineLearnPai2-2.control&depth_1-utm_source=distribute.pc_relevant.none-task-blog-BlogCommendFromMachineLearnPai2-2.control

https://github.com/ljcan/SparkStreaming

發表評論
所有評論
還沒有人評論,想成為第一個評論的人麼? 請在上方評論欄輸入並且點擊發布.
相關文章