1. Data truncation: Data too long for column 'userid' at row 1&&2.Duplicate entry '2020-09-16-??-??' for key 'PRIMARY'

1.Data truncation: Data too long for column 'userid' at row 1具体错误如下:

解决方案:

  You have three choices to fix this.(stackOverflow的解决方法:)

    1. Increase the size of your column datatype.

     2. Decrease the size of the content being inserted into that column.

     3. Don't insert the information.

当处理完这个错误后,发现另一个错误

2.Duplicate entry '2020-09-16-??-??' for key 'PRIMARY'

则不妨:在连接数据库时加入一句话  jdbc:mysql://hadoop102:3306/spark2020?characterEncoding=UTF-8

  再次运行即可

20/09/16 20:43:51 ERROR Executor: Exception in task 0.0 in stage 5.0 (TID 15)
com.mysql.jdbc.MysqlDataTruncation: Data truncation: Data too long for column 'userid' at row 1
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4235)
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4169)
at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:2617)
at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2778)
at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2825)
at com.mysql.jdbc.PreparedStatement.executeInternal(PreparedStatement.java:2156)
at com.mysql.jdbc.PreparedStatement.execute(PreparedStatement.java:1379)
at com.atguigu.bigdata.spark.streaming.SparkStreamingTest12_Req1.anonfunmain8(SparkStreamingTest12_Req1.scala:171)
at com.atguigu.bigdata.spark.streaming.SparkStreamingTest12_Req1.anonfunmain8adapted(SparkStreamingTest12Req1.scala:98)atscala.collection.Iterator.foreach(Iterator.scala:943)atscala.collection.Iterator.foreach(Iterator.scala:943)
at org.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28)
at org.apache.spark.rdd.RDD.anonfunforeach2(RDD.scala:986)atorg.apache.spark.rdd.RDD.anonfunforeach2adapted(RDD.scala:986)atorg.apache.spark.SparkContext.anonfunrunJob5(SparkContext.scala:2133)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:127)
at org.apache.spark.executor.ExecutorTaskRunner.anonfunrun3(Executor.scala:444)
at org.apache.spark.util.Utils.tryWithSafeFinally(Utils.scala:1377)atorg.apache.spark.executor.ExecutorTaskRunner.run(Executor.scala:447)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutorWorker.run(ThreadPoolExecutor.java:624)atjava.lang.Thread.run(Thread.java:748)20/09/1620:43:51ERRORExecutor:Exceptionintask2.0instage5.0(TID16)com.mysql.jdbc.MysqlDataTruncation:Datatruncation:Datatoolongforcolumnuseridatrow1atcom.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4235)atcom.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4169)atcom.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:2617)atcom.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2778)atcom.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2825)atcom.mysql.jdbc.PreparedStatement.executeInternal(PreparedStatement.java:2156)atcom.mysql.jdbc.PreparedStatement.execute(PreparedStatement.java:1379)atcom.atguigu.bigdata.spark.streaming.SparkStreamingTest12Req1.anonfunmain8(SparkStreamingTest12Req1.scala:171)atcom.atguigu.bigdata.spark.streaming.SparkStreamingTest12Req1.anonfunmain8adapted(SparkStreamingTest12_Req1.scala:98)
at scala.collection.Iterator.foreach(Iterator.scala:943)
at scala.collection.Iterator.foreach(Iterator.scala:943)atorg.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28)atorg.apache.spark.rdd.RDD.anonfunforeach2(RDD.scala:986)
at org.apache.spark.rdd.RDD.anonfunforeach2adapted(RDD.scala:986)
at org.apache.spark.SparkContext.anonfunrunJob5(SparkContext.scala:2133)atorg.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)atorg.apache.spark.scheduler.Task.run(Task.scala:127)atorg.apache.spark.executor.ExecutorTaskRunner.anonfunrun3(Executor.scala:444)atorg.apache.spark.util.Utils.tryWithSafeFinally(Utils.scala:1377)
at org.apache.spark.executor.ExecutorTaskRunner.run(Executor.scala:447)atjava.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)atjava.util.concurrent.ThreadPoolExecutorWorker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
20/09/16 20:43:51 ERROR TaskSetManager: Task 0 in stage 5.0 failed 1 times; aborting job
20/09/16 20:43:51 ERROR JobScheduler: Error running job streaming job 1600260231000 ms.0
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 5.0 failed 1 times, most recent failure: Lost task 0.0 in stage 5.0 (TID 15, zls, executor driver): com.mysql.jdbc.MysqlDataTruncation: Data truncation: Data too long for column 'userid' at row 1
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4235)
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4169)
at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:2617)
at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2778)
at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2825)
at com.mysql.jdbc.PreparedStatement.executeInternal(PreparedStatement.java:2156)
at com.mysql.jdbc.PreparedStatement.execute(PreparedStatement.java:1379)
at com.atguigu.bigdata.spark.streaming.SparkStreamingTest12_Req1.anonfunmain8(SparkStreamingTest12_Req1.scala:171)
at com.atguigu.bigdata.spark.streaming.SparkStreamingTest12_Req1.anonfunmain8adapted(SparkStreamingTest12Req1.scala:98)atscala.collection.Iterator.foreach(Iterator.scala:943)atscala.collection.Iterator.foreach(Iterator.scala:943)
at org.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28)
at org.apache.spark.rdd.RDD.anonfunforeach2(RDD.scala:986)atorg.apache.spark.rdd.RDD.anonfunforeach2adapted(RDD.scala:986)atorg.apache.spark.SparkContext.anonfunrunJob5(SparkContext.scala:2133)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:127)
at org.apache.spark.executor.ExecutorTaskRunner.anonfunrun3(Executor.scala:444)
at org.apache.spark.util.Utils.tryWithSafeFinally(Utils.scala:1377)atorg.apache.spark.executor.ExecutorTaskRunner.run(Executor.scala:447)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)

Driver stacktrace:
at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2023)
at org.apache.spark.scheduler.DAGScheduler.anonfunabortStage2(DAGScheduler.scala:1972)atorg.apache.spark.scheduler.DAGScheduler.anonfunabortStage2adapted(DAGScheduler.scala:1971)atscala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)atscala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:55)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1971)
at org.apache.spark.scheduler.DAGScheduler.anonfunhandleTaskSetFailed1(DAGScheduler.scala:950)atorg.apache.spark.scheduler.DAGScheduler.anonfunhandleTaskSetFailed1adapted(DAGScheduler.scala:950)atscala.Option.foreach(Option.scala:407)atorg.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:950)atorg.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2203)atorg.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2152)atorg.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2141)atorg.apache.spark.util.EventLoopanon1.run(EventLoop.scala:49)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:752)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2093)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2114)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2133)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2158)
at org.apache.spark.rdd.RDD.anonfunforeach1(RDD.scala:986)atorg.apache.spark.rdd.RDDOperationScope.withScope(RDDOperationScope.scala:151)
at org.apache.spark.rdd.RDDOperationScope.withScope(RDDOperationScope.scala:112)atorg.apache.spark.rdd.RDD.withScope(RDD.scala:388)atorg.apache.spark.rdd.RDD.foreach(RDD.scala:984)atcom.atguigu.bigdata.spark.streaming.SparkStreamingTest12Req1.anonfunmain7(SparkStreamingTest12Req1.scala:98)atcom.atguigu.bigdata.spark.streaming.SparkStreamingTest12Req1.anonfunmain7adapted(SparkStreamingTest12_Req1.scala:97)
at org.apache.spark.streaming.dstream.DStream.anonfunforeachRDD2(DStream.scala:629)atorg.apache.spark.streaming.dstream.DStream.anonfunforeachRDD2adapted(DStream.scala:629)atorg.apache.spark.streaming.dstream.ForEachDStream.anonfungenerateJob2(ForEachDStream.scala:51)
at scala.runtime.java8.JFunction0mcVsp.apply(JFunction0mcVsp.java:23)
at org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:417)
at org.apache.spark.streaming.dstream.ForEachDStream.anonfungenerateJob1(ForEachDStream.scala:51)atscala.runtime.java8.JFunction0mcVsp.apply(JFunction0mcVsp.java:23)atscala.util.Try.apply(Try.scala:213)
at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39)
at org.apache.spark.streaming.scheduler.JobSchedulerJobHandler.anonfunrun1(JobScheduler.scala:256)
at scala.runtime.java8.JFunction0mcVsp.apply(JFunction0mcVsp.java:23)
at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)
at org.apache.spark.streaming.scheduler.JobSchedulerJobHandler.run(JobScheduler.scala:256)atjava.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)atjava.util.concurrent.ThreadPoolExecutorWorker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: com.mysql.jdbc.MysqlDataTruncation: Data truncation: Data too long for column 'userid' at row 1
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4235)
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4169)
at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:2617)
at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2778)
at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2825)
at com.mysql.jdbc.PreparedStatement.executeInternal(PreparedStatement.java:2156)
at com.mysql.jdbc.PreparedStatement.execute(PreparedStatement.java:1379)
at com.atguigu.bigdata.spark.streaming.SparkStreamingTest12_Req1.anonfunmain8(SparkStreamingTest12_Req1.scala:171)
at com.atguigu.bigdata.spark.streaming.SparkStreamingTest12_Req1.anonfunmain8adapted(SparkStreamingTest12Req1.scala:98)atscala.collection.Iterator.foreach(Iterator.scala:943)atscala.collection.Iterator.foreach(Iterator.scala:943)
at org.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28)
at org.apache.spark.rdd.RDD.anonfunforeach2(RDD.scala:986)atorg.apache.spark.rdd.RDD.anonfunforeach2adapted(RDD.scala:986)atorg.apache.spark.SparkContext.anonfunrunJob5(SparkContext.scala:2133)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:127)
at org.apache.spark.executor.ExecutorTaskRunner.anonfunrun3(Executor.scala:444)
at org.apache.spark.util.Utils.tryWithSafeFinally(Utils.scala:1377)atorg.apache.spark.executor.ExecutorTaskRunner.run(Executor.scala:447)
... 3 more
Exception in thread "main" org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 5.0 failed 1 times, most recent failure: Lost task 0.0 in stage 5.0 (TID 15, zls, executor driver): com.mysql.jdbc.MysqlDataTruncation: Data truncation: Data too long for column 'userid' at row 1
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4235)
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4169)
at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:2617)
at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2778)
at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2825)
at com.mysql.jdbc.PreparedStatement.executeInternal(PreparedStatement.java:2156)
at com.mysql.jdbc.PreparedStatement.execute(PreparedStatement.java:1379)
at com.atguigu.bigdata.spark.streaming.SparkStreamingTest12_Req1.anonfunmain8(SparkStreamingTest12_Req1.scala:171)
at com.atguigu.bigdata.spark.streaming.SparkStreamingTest12_Req1.anonfunmain8adapted(SparkStreamingTest12Req1.scala:98)atscala.collection.Iterator.foreach(Iterator.scala:943)atscala.collection.Iterator.foreach(Iterator.scala:943)
at org.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28)
at org.apache.spark.rdd.RDD.anonfunforeach2(RDD.scala:986)atorg.apache.spark.rdd.RDD.anonfunforeach2adapted(RDD.scala:986)atorg.apache.spark.SparkContext.anonfunrunJob5(SparkContext.scala:2133)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:127)
at org.apache.spark.executor.ExecutorTaskRunner.anonfunrun3(Executor.scala:444)
at org.apache.spark.util.Utils.tryWithSafeFinally(Utils.scala:1377)atorg.apache.spark.executor.ExecutorTaskRunner.run(Executor.scala:447)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)

Driver stacktrace:
at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2023)
at org.apache.spark.scheduler.DAGScheduler.anonfunabortStage2(DAGScheduler.scala:1972)atorg.apache.spark.scheduler.DAGScheduler.anonfunabortStage2adapted(DAGScheduler.scala:1971)atscala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)atscala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:55)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1971)
at org.apache.spark.scheduler.DAGScheduler.anonfunhandleTaskSetFailed1(DAGScheduler.scala:950)atorg.apache.spark.scheduler.DAGScheduler.anonfunhandleTaskSetFailed1adapted(DAGScheduler.scala:950)atscala.Option.foreach(Option.scala:407)atorg.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:950)atorg.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2203)atorg.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2152)atorg.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2141)atorg.apache.spark.util.EventLoopanon1.run(EventLoop.scala:49)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:752)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2093)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2114)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2133)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2158)
at org.apache.spark.rdd.RDD.anonfunforeach1(RDD.scala:986)atorg.apache.spark.rdd.RDDOperationScope.withScope(RDDOperationScope.scala:151)
at org.apache.spark.rdd.RDDOperationScope.withScope(RDDOperationScope.scala:112)atorg.apache.spark.rdd.RDD.withScope(RDD.scala:388)atorg.apache.spark.rdd.RDD.foreach(RDD.scala:984)atcom.atguigu.bigdata.spark.streaming.SparkStreamingTest12Req1.anonfunmain7(SparkStreamingTest12Req1.scala:98)atcom.atguigu.bigdata.spark.streaming.SparkStreamingTest12Req1.anonfunmain7adapted(SparkStreamingTest12_Req1.scala:97)
at org.apache.spark.streaming.dstream.DStream.anonfunforeachRDD2(DStream.scala:629)atorg.apache.spark.streaming.dstream.DStream.anonfunforeachRDD2adapted(DStream.scala:629)atorg.apache.spark.streaming.dstream.ForEachDStream.anonfungenerateJob2(ForEachDStream.scala:51)
at scala.runtime.java8.JFunction0mcVsp.apply(JFunction0mcVsp.java:23)
at org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:417)
at org.apache.spark.streaming.dstream.ForEachDStream.anonfungenerateJob1(ForEachDStream.scala:51)atscala.runtime.java8.JFunction0mcVsp.apply(JFunction0mcVsp.java:23)atscala.util.Try.apply(Try.scala:213)
at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39)
at org.apache.spark.streaming.scheduler.JobSchedulerJobHandler.anonfunrun1(JobScheduler.scala:256)
at scala.runtime.java8.JFunction0mcVsp.apply(JFunction0mcVsp.java:23)
at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)
at org.apache.spark.streaming.scheduler.JobSchedulerJobHandler.run(JobScheduler.scala:256)atjava.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)atjava.util.concurrent.ThreadPoolExecutorWorker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: com.mysql.jdbc.MysqlDataTruncation: Data truncation: Data too long for column 'userid' at row 1
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4235)
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4169)
at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:2617)
at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2778)
at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2825)
at com.mysql.jdbc.PreparedStatement.executeInternal(PreparedStatement.java:2156)
at com.mysql.jdbc.PreparedStatement.execute(PreparedStatement.java:1379)
at com.atguigu.bigdata.spark.streaming.SparkStreamingTest12_Req1.anonfunmain8(SparkStreamingTest12_Req1.scala:171)
at com.atguigu.bigdata.spark.streaming.SparkStreamingTest12_Req1.anonfunmain8adapted(SparkStreamingTest12Req1.scala:98)atscala.collection.Iterator.foreach(Iterator.scala:943)atscala.collection.Iterator.foreach(Iterator.scala:943)
at org.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28)
at org.apache.spark.rdd.RDD.anonfunforeach2(RDD.scala:986)atorg.apache.spark.rdd.RDD.anonfunforeach2adapted(RDD.scala:986)atorg.apache.spark.SparkContext.anonfunrunJob5(SparkContext.scala:2133)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:127)
at org.apache.spark.executor.ExecutorTaskRunner.anonfunrun3(Executor.scala:444)
at org.apache.spark.util.Utils.tryWithSafeFinally(Utils.scala:1377)atorg.apache.spark.executor.ExecutorTaskRunner.run(Executor.scala:447)
... 3 more

Process finished with exit code 1

posted @   Mrzxs  阅读(310)  评论(0编辑  收藏  举报
编辑推荐:
· AI与.NET技术实操系列(二):开始使用ML.NET
· 记一次.NET内存居高不下排查解决与启示
· 探究高空视频全景AR技术的实现原理
· 理解Rust引用及其生命周期标识(上)
· 浏览器原生「磁吸」效果!Anchor Positioning 锚点定位神器解析
阅读排行:
· DeepSeek 开源周回顾「GitHub 热点速览」
· 物流快递公司核心技术能力-地址解析分单基础技术分享
· .NET 10首个预览版发布:重大改进与新特性概览!
· AI与.NET技术实操系列(二):开始使用ML.NET
· 单线程的Redis速度为什么快?
点击右上角即可分享
微信分享提示
more_horiz
keyboard_arrow_up dark_mode palette
选择主题