Java自学者论坛

 找回密码
 立即注册

手机号码,快捷登录

恭喜Java自学者论坛(https://www.javazxz.com)已经为数万Java学习者服务超过8年了!积累会员资料超过10000G+
成为本站VIP会员,下载本站10000G+会员资源,会员资料板块,购买链接:点击进入购买VIP会员

JAVA高级面试进阶训练营视频教程

Java架构师系统进阶VIP课程

分布式高可用全栈开发微服务教程Go语言视频零基础入门到精通Java架构师3期(课件+源码)
Java开发全终端实战租房项目视频教程SpringBoot2.X入门到高级使用教程大数据培训第六期全套视频教程深度学习(CNN RNN GAN)算法原理Java亿级流量电商系统视频教程
互联网架构师视频教程年薪50万Spark2.0从入门到精通年薪50万!人工智能学习路线教程年薪50万大数据入门到精通学习路线年薪50万机器学习入门到精通教程
仿小米商城类app和小程序视频教程深度学习数据分析基础到实战最新黑马javaEE2.1就业课程从 0到JVM实战高手教程MySQL入门到精通教程
查看: 843|回复: 0

Hadoop:开发机运行spark程序,抛出异常:ERROR Shell: Failed to locate the winutils binary in the hadoop binary path

[复制链接]
  • TA的每日心情
    奋斗
    2024-4-6 11:05
  • 签到天数: 748 天

    [LV.9]以坛为家II

    2034

    主题

    2092

    帖子

    70万

    积分

    管理员

    Rank: 9Rank: 9Rank: 9

    积分
    705612
    发表于 2021-6-7 01:37:30 | 显示全部楼层 |阅读模式

    问题:

    windows开发机运行spark程序,抛出异常:ERROR Shell: Failed to locate the winutils binary in the hadoop binary path,但是可以正常执行,并不影响结果。

    18/07/02 19:46:08 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
    18/07/02 19:46:08 ERROR Shell: Failed to locate the winutils binary in the hadoop binary path
    java.io.IOException: Could not locate executable null\bin\winutils.exe in the Hadoop binaries.
        at org.apache.hadoop.util.Shell.getQualifiedBinPath(Shell.java:355)
        at org.apache.hadoop.util.Shell.getWinUtilsPath(Shell.java:370)
        at org.apache.hadoop.util.Shell.<clinit>(Shell.java:363)
        at org.apache.hadoop.util.StringUtils.<clinit>(StringUtils.java:79)
        at org.apache.hadoop.security.Groups.parseStaticMapping(Groups.java:116)
        at org.apache.hadoop.security.Groups.<init>(Groups.java:93)
        at org.apache.hadoop.security.Groups.<init>(Groups.java:73)
        at org.apache.hadoop.security.Groups.getUserToGroupsMappingService(Groups.java:293)
        at org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:283)
        at org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:260)
        at org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:789)
        at org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:774)
        at org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:647)
        at org.apache.spark.util.Utils$$anonfun$getCurrentUserName$1.apply(Utils.scala:2198)
        at org.apache.spark.util.Utils$$anonfun$getCurrentUserName$1.apply(Utils.scala:2198)
        at scala.Option.getOrElse(Option.scala:120)
        at org.apache.spark.util.Utils$.getCurrentUserName(Utils.scala:2198)
        at org.apache.spark.SparkContext.<init>(SparkContext.scala:322)
        at org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:59)
        at com.lm.sparkLearning.utils.SparkUtils.getJavaSparkContext(SparkUtils.java:31)
        at com.lm.sparkLearning.rdd.RddLearning.main(RddLearning.java:30)
    18/07/02 19:46:14 WARN RddLearning: singleOperateRdd mapRdd->[2, 3, 4, 4]
    18/07/02 19:46:14 WARN RddLearning: singleOperateRdd flatMapRdd->[2, 3, 2, 3, 2, 3, 2, 3]
    18/07/02 19:46:14 WARN RddLearning: singleOperateRdd filterRdd->[3, 3]
    18/07/02 19:46:14 WARN RddLearning: singleOperateRdd distinctRdd->[2, 1, 3]
    18/07/02 19:46:14 WARN RddLearning: singleOperateRdd sampleRdd->[1, 3]
    18/07/02 19:46:14 WARN RddLearning: the program end

    这里所执行的程序是:

    package com.lm.sparkLearning.rdd;
    
    import java.util.Arrays;
    import java.util.Iterator;
    import java.util.List;
    
    import org.apache.spark.api.java.JavaPairRDD;
    import org.apache.spark.api.java.JavaRDD;
    import org.apache.spark.api.java.JavaSparkContext;
    import org.apache.spark.api.java.function.FlatMapFunction;
    import org.apache.spark.api.java.function.Function;
    import org.apache.spark.api.java.function.Function2;
    import org.apache.spark.api.java.function.VoidFunction;
    import org.slf4j.Logger;
    import org.slf4j.LoggerFactory;
    
    import com.lm.sparkLearning.utils.SparkUtils;
    
    public class RddLearning {
        private static Logger logger = LoggerFactory.getLogger(RddLearning.class);
    
        public static void main(String[] args) {
    
            JavaSparkContext jsc = SparkUtils.getJavaSparkContext("RDDLearning", "local[2]", "WARN");
    
            SparkUtils.createRddExternal(jsc, "D:/README.txt");
            singleOperateRdd(jsc);
    
            jsc.stop();
            
            logger.warn("the program end");
        }
    
        public static void singleOperateRdd(JavaSparkContext jsc) {
            List<Integer> nums = Arrays.asList(new Integer[] { 1, 2, 3, 3 });
            JavaRDD<Integer> numsRdd = SparkUtils.createRddCollect(jsc, nums);
    
            // map
            JavaRDD<Integer> mapRdd = numsRdd.map(new Function<Integer, Integer>() {
                private static final long serialVersionUID = 1L;
    
                @Override
                public Integer call(Integer v1) throws Exception {
                    return (v1 + 1);
                }
            });
    
            logger.warn("singleOperateRdd mapRdd->" + mapRdd.collect().toString());
    
            JavaRDD<Integer> flatMapRdd = numsRdd.flatMap(new FlatMapFunction<Integer, Integer>() {
                private static final long serialVersionUID = 1L;
    
                @Override
                public Iterable<Integer> call(Integer t) throws Exception {
                    return Arrays.asList(new Integer[] { 2, 3 });
                }
            });
    
            logger.warn("singleOperateRdd flatMapRdd->" + flatMapRdd.collect().toString());
    
            JavaRDD<Integer> filterRdd = numsRdd.filter(new Function<Integer, Boolean>() {
                private static final long serialVersionUID = 1L;
    
                @Override
                public Boolean call(Integer v1) throws Exception {
                    return v1 > 2;
                }
            });
    
            logger.warn("singleOperateRdd filterRdd->" + filterRdd.collect().toString());
    
            JavaRDD<Integer> distinctRdd = numsRdd.distinct();
    
            logger.warn("singleOperateRdd distinctRdd->" + distinctRdd.collect().toString());
    
            JavaRDD<Integer> sampleRdd = numsRdd.sample(false, 0.5);
    
            logger.warn("singleOperateRdd sampleRdd->" + sampleRdd.collect().toString());
        }
    }

    解决方案:

    1.下载winutils的windows版本
    GitHub上,有人提供了winutils的windows的版本,项目地址是:https://github.com/srccodes/hadoop-common-2.2.0-bin,直接下载此项目的zip包,下载后是文件名是hadoop-common-2.2.0-bin-master.zip,随便解压到一个目录。
    2.配置环境变量
    增加用户变量HADOOP_HOME,值是下载的zip包解压的目录,然后在系统变量path里增加$HADOOP_HOME\bin 即可。

    添加“%HADOOP%\bin”到path


    再次运行程序,正常执行。

     

    哎...今天够累的,签到来了1...
    回复

    使用道具 举报

    您需要登录后才可以回帖 登录 | 立即注册

    本版积分规则

    QQ|手机版|小黑屋|Java自学者论坛 ( 声明:本站文章及资料整理自互联网,用于Java自学者交流学习使用,对资料版权不负任何法律责任,若有侵权请及时联系客服屏蔽删除 )

    GMT+8, 2024-4-30 20:03 , Processed in 0.062653 second(s), 29 queries .

    Powered by Discuz! X3.4

    Copyright © 2001-2021, Tencent Cloud.

    快速回复 返回顶部 返回列表