Skip to main content
Stack Overflow
  1. About
  2. For Teams
Filter by
Sorted by
Tagged with
-1 votes
2 answers
54 views

I must install HBase on my Virtual Machines, which are Linux Ubuntu LTS 20.04. I was wondering if there was a way to determine which HBase version is compatible with, because I saw the mention of ...
0 votes
0 answers
12 views

When I run start-dfs.sh, I don't see the NameNode specified in the command's output. I have written the following in the hdfs-site.xml document: <?xml version="1.0" encoding="UTF-8&...
1 vote
0 answers
53 views

I understand that Journal Nodes are like the central repo for all edit logs (no matter which namenode is currently active, all push to journal node). I suppose the QuorumJournalManager handles this ...
0 votes
1 answer
91 views

The goal I want to switch from the HDFS to the s3a client. To this end I need to upgrade from Hadoop 2.8.5 to at least 3.1.2, because I need to use the AssumedRoleCredentialProvider for AWS access. ...
1 vote
0 answers
82 views

We are using Hadoop version 2.10.2 and facing below error while starting server. Due to company security policy no execution permission is set on /tmp directory hence library libleveldbjni-64-1-...
1 vote
1 answer
2k views

I have trino to query hdfs with hive connector. not always but sometimes it gets this error : io.trino.spi.trinoexception error reading from hdfs at position caused by java.io.ioexception 4 missing ...
2 votes
1 answer
423 views

I need to configure the value of hadoop.service.shutdown.timeout due to the shutdown hooks triggering a timeout when our MR jobs stop: 2023年08月25日 08:44:39,566 [WARN] [Thread-0] [org.apache.hadoop.util....
0 votes
1 answer
155 views

I'm trying to set up an Accumulo cluster that uses a separate ZooKeeper cluster. I've configured the accumulo-site.xml file to include the instance.zookeeper.host property with the hostname or IP ...
1 vote
0 answers
184 views

I am trying to run Junits for a spark project in intelliji. Junits initialize local hadoop cluster using hadoop-minicluster dependency. Tests run fine with hadoop version - 2.7.3.2.6.5.0-292. Since we ...
mbr's user avatar
  • 21
0 votes
1 answer
675 views

So I have a project directory "dataplatform" and its contents are the follows: ── dataplatform ├── __init__.py ├── commons │ ├── __init__.py │ ├── ...
0 votes
1 answer
185 views

I get get running apps with this yarn application -appStates RUNNING then I get one applicationID from list. then I can get status of app with this: yarn application -status I want to get job ...
CompEng's user avatar
  • 7,426
-1 votes
1 answer
182 views

How can I get the value of one or more keys in HDFS via HTTP or JAVA api from remote client? For example, the file below has a million keys and values. I just want to get the values of the 'phone' and ...
0 votes
0 answers
414 views

Trying on the wordcount example by using command "hadoop jar share/hadoop/mapreduce/hadoop-mapreduce-examples-3.1.3.jar wordcount /input /output" in linux shell, however it keep reminds me ...
1 vote
2 answers
613 views

I am trying to understand the difference between FileInputStream Vs FSDataInputStream and FileOutputStream Vs FSDataOutputStream. I am trying to read a file from S3 bucket and apply some formatting ...
Babu's user avatar
  • 169
1 vote
1 answer
1k views

Below is the sample code snippet that is used for data fetch from HBase. This worked fine with Spark 3.1.2. However after upgrading to Spark 3.2.1, it is not working i.e. returned RDD doesn't contain ...

15 30 50 per page
1
2 3 4 5
...
138

AltStyle によって変換されたページ (->オリジナル) /