<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: org.apache.spark.SparkContext - Error initializing SparkContext in Talend Studio</title>
    <link>https://community.qlik.com/t5/Talend-Studio/org-apache-spark-SparkContext-Error-initializing-SparkContext/m-p/2230330#M20981</link>
    <description>&lt;P&gt;Hello,&lt;/P&gt; 
&lt;P&gt;Are you using jdk 1.8? Is your&amp;nbsp;cluster correctly configured and your connection from the repository? More information will be preferred.&lt;/P&gt; 
&lt;P&gt;Best regards&lt;/P&gt; 
&lt;P&gt;Sabrina&lt;/P&gt; 
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Thu, 11 Oct 2018 04:25:51 GMT</pubDate>
    <dc:creator>Anonymous</dc:creator>
    <dc:date>2018-10-11T04:25:51Z</dc:date>
    <item>
      <title>org.apache.spark.SparkContext - Error initializing SparkContext</title>
      <link>https://community.qlik.com/t5/Talend-Studio/org-apache-spark-SparkContext-Error-initializing-SparkContext/m-p/2230329#M20980</link>
      <description>&lt;P&gt;hi,&lt;/P&gt; 
&lt;P&gt;&amp;nbsp;&lt;/P&gt; 
&lt;P&gt;i'm new to talend big data integration,&lt;BR /&gt;i'm currently try to create spark big data batch job in talend, and&amp;nbsp; encounter following error.&lt;/P&gt; 
&lt;P&gt;the job is only to read data from hive, using talend big data batch to tlogrow&lt;BR /&gt;&lt;SPAN class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="talend big data batch.png" style="width: 999px;"&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="0683p000009M076.png"&gt;&lt;img src="https://community.qlik.com/t5/image/serverpage/image-id/151032i26470B5882436130/image-size/large?v=v2&amp;amp;px=999" role="button" title="0683p000009M076.png" alt="0683p000009M076.png" /&gt;&lt;/span&gt;&lt;/SPAN&gt;&lt;BR /&gt;Starting job TEST_SPARK at 15:23 30/09/2018.&lt;BR /&gt;&lt;BR /&gt;[statistics] connecting to socket on port 3746&lt;BR /&gt;[statistics] connected&lt;BR /&gt;SLF4J: Class path contains multiple SLF4J bindings.&lt;BR /&gt;SLF4J: Found binding in [jar:file:/D:/TalendStudio/talendworkspace/.Java/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]&lt;BR /&gt;SLF4J: Found binding in [jar:file:/D:/TalendStudio/talendworkspace/.Java/lib/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]&lt;BR /&gt;SLF4J: See &lt;A href="http://www.slf4j.org/codes.html#multiple_bindings" target="_blank" rel="nofollow noopener noreferrer"&gt;http://www.slf4j.org/codes.html#multiple_bindings&lt;/A&gt; for an explanation.&lt;BR /&gt;SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]&lt;BR /&gt;[WARN ]: org.apache.hadoop.util.NativeCodeLoader - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable&lt;BR /&gt;[WARN ]: org.apache.spark.SparkConf - In Spark 1.0 and later spark.local.dir will be overridden by the value set by the cluster manager (via SPARK_LOCAL_DIRS in mesos/standalone and LOCAL_DIRS in YARN).&lt;BR /&gt;[WARN ]: org.apache.hadoop.hdfs.shortcircuit.DomainSocketFactory - The short-circuit local reads feature cannot be used because UNIX Domain sockets are not available on Windows.&lt;BR /&gt;[WARN ]: org.apache.hadoop.hdfs.DFSClient - DFSOutputStream ResponseProcessor exception&amp;nbsp; for block BP-1971060428-10.1.98.58-1536015946021:blk_1073769294_28491&lt;BR /&gt;java.io.IOException: An existing connection was forcibly closed by the remote host&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at sun.nio.ch.SocketDispatcher.read0(Native Method)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:43)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at sun.nio.ch.IOUtil.read(IOUtil.java:197)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:380)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:57)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:118)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at java.io.FilterInputStream.read(FilterInputStream.java:83)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at java.io.FilterInputStream.read(FilterInputStream.java:83)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.protocolPB.PBHelper.vintPrefixed(PBHelper.java:2390)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer$ResponseProcessor.run(DFSOutputStream.java:837)&lt;BR /&gt;[WARN ]: org.apache.hadoop.hdfs.DFSClient - Error Recovery for block BP-1971060428-10.1.98.58-1536015946021:blk_1073769294_28491 in pipeline DatanodeInfoWithStorage[10.1.98.60:50010,DS-c101ddeb-7b7b-4bce-a539-56779c4d2787,DISK], DatanodeInfoWithStorage[10.1.98.61:50010,DS-a1a58bf4-8cb2-4795-a8ac-b06bf9160196,DISK]: bad datanode DatanodeInfoWithStorage[10.1.98.60:50010,DS-c101ddeb-7b7b-4bce-a539-56779c4d2787,DISK]&lt;BR /&gt;[WARN ]: org.apache.hadoop.hdfs.DFSClient - DataStreamer Exception&lt;BR /&gt;java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[DatanodeInfoWithStorage[10.1.98.61:50010,DS-a1a58bf4-8cb2-4795-a8ac-b06bf9160196,DISK]], original=[DatanodeInfoWithStorage[10.1.98.61:50010,DS-a1a58bf4-8cb2-4795-a8ac-b06bf9160196,DISK]]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration.&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.findNewDatanode(DFSOutputStream.java:1036)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1110)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1268)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.processDatanodeError(DFSOutputStream.java:993)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:500)&lt;BR /&gt;[ERROR]: org.apache.spark.SparkContext - Error initializing SparkContext.&lt;BR /&gt;java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[DatanodeInfoWithStorage[10.1.98.61:50010,DS-a1a58bf4-8cb2-4795-a8ac-b06bf9160196,DISK]], original=[DatanodeInfoWithStorage[10.1.98.61:50010,DS-a1a58bf4-8cb2-4795-a8ac-b06bf9160196,DISK]]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration.&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.findNewDatanode(DFSOutputStream.java:1036)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1110)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1268)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.processDatanodeError(DFSOutputStream.java:993)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:500)&lt;BR /&gt;[WARN ]: org.apache.spark.scheduler.cluster.YarnSchedulerBackend$YarnSchedulerEndpoint - Attempted to request executors before the AM has registered!&lt;BR /&gt;[WARN ]: org.apache.spark.metrics.MetricsSystem - Stopping a MetricsSystem that is not running&lt;BR /&gt;java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[DatanodeInfoWithStorage[10.1.98.61:50010,DS-a1a58bf4-8cb2-4795-a8ac-b06bf9160196,DISK]], original=[DatanodeInfoWithStorage[10.1.98.61:50010,DS-a1a58bf4-8cb2-4795-a8ac-b06bf9160196,DISK]]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration.&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.findNewDatanode(DFSOutputStream.java:1036)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1110)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1268)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.processDatanodeError(DFSOutputStream.java:993)&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:500)&lt;BR /&gt;[ERROR]: eds_spark_test.test_spark_0_1.TEST_SPARK - TalendJob: 'TEST_SPARK' - Failed with exit code: 1.&lt;BR /&gt;Exception in thread "main" java.lang.RuntimeException: TalendJob: 'TEST_SPARK' - Failed with exit code: 1.&lt;BR /&gt;&amp;nbsp;&amp;nbsp; &amp;nbsp;at eds_spark_test.test_spark_0_1.TEST_SPARK.main(TEST_SPARK.java:1049)&lt;BR /&gt;Job TEST_SPARK ended at 15:31 30/09/2018. [exit code=0]&lt;/P&gt; 
&lt;P&gt;&amp;nbsp;&lt;/P&gt; 
&lt;P&gt;&amp;nbsp;&lt;/P&gt; 
&lt;P&gt;i'm not sure whats causing this.&lt;/P&gt; 
&lt;P&gt;i'm currently using,&lt;/P&gt; 
&lt;P&gt;Talend Big Data 6.4.1 and Hortonworks HDP 2.6.5.0&lt;/P&gt; 
&lt;P&gt;let me know if more detail needed.&lt;/P&gt;</description>
      <pubDate>Sat, 16 Nov 2024 07:33:17 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/org-apache-spark-SparkContext-Error-initializing-SparkContext/m-p/2230329#M20980</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2024-11-16T07:33:17Z</dc:date>
    </item>
    <item>
      <title>Re: org.apache.spark.SparkContext - Error initializing SparkContext</title>
      <link>https://community.qlik.com/t5/Talend-Studio/org-apache-spark-SparkContext-Error-initializing-SparkContext/m-p/2230330#M20981</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt; 
&lt;P&gt;Are you using jdk 1.8? Is your&amp;nbsp;cluster correctly configured and your connection from the repository? More information will be preferred.&lt;/P&gt; 
&lt;P&gt;Best regards&lt;/P&gt; 
&lt;P&gt;Sabrina&lt;/P&gt; 
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 11 Oct 2018 04:25:51 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/org-apache-spark-SparkContext-Error-initializing-SparkContext/m-p/2230330#M20981</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2018-10-11T04:25:51Z</dc:date>
    </item>
  </channel>
</rss>

