展开
拓展 关闭
订阅号推广码
GitHub
视频
公告栏 关闭

spark常见错误

  • 查看spark默认的python版本
[root@master day27]# pyspark
/home/software/spark-2.3.4-bin-hadoop2.7/conf/spark-env.sh: line 2: /usr/local/hadoop/bin/hadoop: No such file or directory
Python 2.7.5 (default, Nov 16 2020, 22:23:17)
[GCC 4.8.5 20150623 (Red Hat 4.8.5-44)] on linux2
Type "help", "copyright", "credits" or "license" for more information.
2027-11-29 18:45:59 WARN NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
Welcome to
____ __
/ __/__ ___ _____/ /__
_\ \/ _ \/ _ `/ __/ '_/
/__ / .__/\_,_/_/ /_/\_\ version 2.3.4
/_/
Using Python version 2.7.5 (default, Nov 16 2020 22:23:17)
SparkSession available as 'spark'.
  • 修改为python3
# 查找位置
[root@master ~]# whereis pyspark
pyspark: /home/software/spark-2.3.4-bin-hadoop2.7/bin/pyspark2.cmd /home/software/spark-2.3.4-bin-hadoop2.7/bin/pyspark.cmd /home/software/spark-2.3.4-bin-hadoop2.7/bin/pyspark
# 进入指定路径
cd /home/software/spark-2.3.4-bin-hadoop2.7/bin
# 编辑
vim pyspark
# python3的路径
export PYSPARK_HOME=/home/software/anaconda3/bin/python3
export PYSPARK_PYTHON=/home/software/anaconda3/bin/python3
  • 报错如下
ssh://root@192.168.128.78:22/home/software/anaconda3/bin/python3 -u /tmp/pycharm_project_115/day27/test1.py
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
27/11/29 18:41:40 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
/home/software/anaconda3/lib/python3.6/site-packages/pyspark/context.py:238: FutureWarning: Python 3.6 support is deprecated in Spark 3.2.
FutureWarning
27/11/29 18:41:49 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID 0)
java.io.IOException: Cannot run program "python3": error=2, No such file or directory
at java.lang.ProcessBuilder.start(ProcessBuilder.java:1048)
at org.apache.spark.api.python.PythonWorkerFactory.startDaemon(PythonWorkerFactory.scala:215)
  • 解决方案
[root@master ~]# cd ~
[root@master ~]# vim .bashrc
# 配置python3的路径
export PYSPARK_HOME=/home/software/anaconda3/bin/python3
export PYSPARK_PYTHON=/home/software/anaconda3/bin/python3
# 配置生效
[root@master ~]# source .bashrc
posted @   DogLeftover  阅读(116)  评论(0编辑  收藏  举报
相关博文:
阅读排行:
· 阿里巴巴 QwQ-32B真的超越了 DeepSeek R-1吗?
· 10年+ .NET Coder 心语 ── 封装的思维:从隐藏、稳定开始理解其本质意义
· 【设计模式】告别冗长if-else语句:使用策略模式优化代码结构
· 字符编码:从基础到乱码解决
· 提示词工程——AI应用必不可少的技术
历史上的今天:
2021-11-30 vue导出pdf
2021-11-30 发送邮件
2021-11-30 文件上传下载
2021-11-30 Mybatis plus案例
2021-11-30 Mybatis plus代码生成器
点击右上角即可分享
微信分享提示