我正在使用Docker Toolbox在Windows 10上使用Spark。
泊坞窗映像为sequenceiq/spark
。 https://hub.docker.com/r/sequenceiq/spark/
但是,会显示很多“日志”。
scala> dataRDD.collect()
18/12/07 03:59:48 INFO mapred.FileInputFormat: Total input paths to process : 1
18/12/07 03:59:49 INFO spark.SparkContext: Starting job: collect at <console>:32
18/12/07 03:59:49 INFO scheduler.DAGScheduler: Got job 0 (collect at <console>:32) with 1 output partitions
18/12/07 03:59:49 INFO scheduler.DAGScheduler: Final stage: ResultStage 0 (collect at <console>:32)
18/12/07 03:59:49 INFO scheduler.DAGScheduler: Parents of final stage: List()
18/12/07 03:59:49 INFO scheduler.DAGScheduler: Missing parents: List()
18/12/07 03:59:49 INFO scheduler.DAGScheduler: Submitting ResultStage 0 (MapPartitionsRDD[5] at map at <console>:29), which has no missing parents
18/12/07 03:59:49 INFO storage.MemoryStore: Block broadcast_2 stored as values in memory (estimated size 3.2 KB, free 357.5 KB)
18/12/07 03:59:49 INFO storage.MemoryStore: Block broadcast_2_piece0 stored as bytes in memory (estimated size 1870.0 B, free 359.3 KB)
18/12/07 03:59:49 INFO storage.BlockManagerInfo: Added broadcast_2_piece0 in memory on localhost:39093 (size: 1870.0 B, free: 517.4 MB)
18/12/07 03:59:49 INFO spark.SparkContext: Created broadcast 2 from broadcast at DAGScheduler.scala:1006
18/12/07 03:59:49 INFO scheduler.DAGScheduler: Submitting 1 missing tasks from ResultStage 0 (MapPartitionsRDD[5] at map at <console>:29)
...
在启动过程中是否可以通过选项隐藏它们?
谢谢
答案 0 :(得分:1)
也许简单
docker logs | grep -V INFO | grep -v WARN
?