Skip to main content
Stack Overflow
  1. About
  2. For Teams
Filter by
Sorted by
Tagged with
0 votes
0 answers
25 views

Setup explanation: I am using jmx prometheus agent, which starts with the VM as vm argument in HDFS setup. Version: 0.20.0 There is a jmx http port displaying the Mbeans as JSON. Problem: When I view ...
0 votes
0 answers
98 views

I have a Flink ETL job that reads from ~13 Kafka topics and writes data into HDFS using a FileSink with compaction enabled. Right now, we have around 40 different output paths (buckets), and roughly ...
2 votes
0 answers
40 views

I have hadoop + hive setup using docker, however when I try to run count(*) on my table it gives me an empty return when using Tez and the correct one when using MapReduce, the table is an external ...
0 votes
0 answers
20 views

I have a 3 node hadoop cluster (version 3.4.1) with java_home pointing to version 8 on each node. I want to evenly distribute the uploaded data across all nodes when I type the following: hdfs ...
vinhdiesal's user avatar
0 votes
1 answer
53 views

I'm currently learning Hadoop and am a bit confused about how the Hadoop Distributed File System (HDFS) and YARN components interact, especially in terms of deployment across machines. Here’s what I ...
0 votes
0 answers
174 views

I am working with Mlflow==2.19.0 in a Red Hat Enterprise Linux Server release 7.9 (Maipo). Everythig works fine except with the log_image method that for some reason is converting parts of the string ...
1 vote
0 answers
95 views

I would like to create file transformation pipelines by downloading input files from remote HDFS and storing outputs files on the same remote HDFS. Kerberos is used to authenticate to my hadoop ...
0 votes
0 answers
27 views

I have a Cloudera cluster under knox gateway with basic authentication (username/password). I want to access HDFS (SWebHDFS) through SSL (https) using Java Apache Hadoop client library (Apache Hadoop ...
0 votes
2 answers
260 views

I am trying to create a new connection from DBeaver to a Delta Lake Parquet file which is located on the HDFS filesystem which I successfully created with a Spark/Hadoop/Scala/io.delta application. (...
0 votes
1 answer
82 views

I'm running flink job and on my local machine I dont see any issue of streaming the data to Azure blob, but when I deploy on dev environment I'm seeing an error in the console like Caused by: org....
0 votes
1 answer
69 views

I have a large number of files (tens of thousands) in a Unix directory that I need to copy to Hadoop using the command: hdfs dfs -put * /hdfs_folder/ However, some of these files have spaces in their ...
1 vote
1 answer
107 views

In Jupyter notebooks I suceed in reading parquet files in HDFS thanks to sparkmagic. Spark magic conf starts with : { "kernel_python_credentials" : { "username": "admin&...
0 votes
1 answer
26 views

Why hadoop node had a nodename like this: iZib208xfvbhmyx1rha3gqZ on an alicloud ECS [root@worker1 hadoop-3.4.1]# hdfs namenode -format 2025年01月23日 10:13:46,887 INFO namenode.NameNode: STARTUP_MSG: /*...
yizhen liu's user avatar
0 votes
1 answer
148 views

Objective : we need to read the table in spark application and trasform the data and rewrite the same table Senario : I am trying to overwrite external non partitioned table with spark Since same data ...
0 votes
0 answers
28 views

Background : HBase reads seem to fall under the 'strong consistency' model, as : All reads are served from the master where the data has already been committed. As a result, the clients seem to always ...

15 30 50 per page
1
2 3 4 5
...
551

AltStyle によって変換されたページ (->オリジナル) /