171 questions
- Bountied 0
- Unanswered
- Frequent
- Score
- Trending
- Week
- Month
- Unanswered (my tags)
0
votes
1
answer
33
views
Apache Flink: I/O writing thread encountered an error: segment has been freed
We have an Flink job (batch mode) that runs on AWS KDA (ver: 1.20.0), where its logical operators look like: FileSource -> map() -> AssignTimestamps() -> filter() -> keyBy -> ...
-4
votes
1
answer
100
views
Questions about Apache Flink internals, BATCH execution mode
We recently experimented with Flink, in particular BATCH execution mode to setup an ETL job processing an bounded data-set. It works quite well but I'd like to get some clarifications about my ...
0
votes
1
answer
65
views
How to limit single file size when using Flink batch mode to write Parquet
I was using Flink in batch mode to read data from one source and then directly write the data into file system as Parquet format.
The code was like:
hudi_source_ddl = f"""
...
0
votes
0
answers
31
views
Issue with Flink Application Mode args – Custom Command Not Executing as Expected
Issue with Flink Application Mode args – Custom Command Not Executing as Expected
Description:
Apache Flink Version: 1.20.1
Java Version: 17
Platform: AWS EKS (Kubernetes)
I'm encountering an issue ...
0
votes
1
answer
165
views
Improperly specified VM option 'MaxMetaspaceSize=268435456 '
While Running Flink on Microsoft Windows 11 Enterprise with Java 17 using flink 2.0.0 getting below error
Improperly specified VM option 'MaxMetaspaceSize=268435456
'
Error: Could not create the Java ...
1
vote
0
answers
92
views
Get Exception after submit the pyFlink Job
I am a new for pyflink. I try to run submit a simple python to YARN application mode. But I got the error said cannot find the python file word_count.py. Below is my environment and the exception log. ...
1
vote
0
answers
52
views
Submitting multiple DELETE statements in Flink SQL
I' m trying to submit multiple DELETE statements in to the TableEnvironment in way:
val settings = EnvironmentSettings.newInstance.inBatchMode.build()
val env = TableEnvironment.create(settings)
...
0
votes
1
answer
86
views
Batch elements in Apache Flink
I have a stream of IDs in Apache Flink. I would like to batch them into sets of 500 and for each batch call an external service that will give me additional data for each ID. Then I want to forward ...
0
votes
1
answer
147
views
Ephemeral storage bloat when submitting multiple jobs to flink session cluster
I’m try to run a flink session cluster and submit a lot of small jobs on it.
My jar is configurable on what processing should happen, where the data is and to where the processed data should be loaded....
0
votes
0
answers
116
views
Command failed with error 18 (AuthenticationFailed): 'Authentication failed.'
Even after providing the correct credentials in the flinkSql code still getting the authentication exception. Please find the exception below. Have tried the python code to connect to the mongo db and ...
0
votes
2
answers
64
views
How do you scale out a flink process to use more processes?
I have a batch job that involves a set of map and reduce steps that I've successfully converted to an Apache Flink application.
As a proof of concept, I deployed this via docker-compose using 3 ...
1
vote
0
answers
94
views
Flink batch state handover
I have a Flink 1.14 backfill batch job large enough that it's very
resource-intensive and hard to run to completion without spurious failures (network glitches, node scheduling failures, disk capacity ...
0
votes
1
answer
315
views
Unable to get CsvReaderFormat for Flink read a CSV file
I am trying a simple task of reading a CSV file inside flink execution class using CsvReaderFormat as mentioned in the documentation. I have a pojo called subscriberADSR (I know it is a bad practise ...
0
votes
1
answer
983
views
Flink Operator stuck at 100% busy, how do I get it down? [closed]
I've deployed a Flink cluster as a yarn application. As part of the yarn configuration I associated 32 vCores to each Task Manager. I also allocated 2 slots to each Task Manager.
Job Pipeline: Kafka ...
0
votes
1
answer
118
views
Flink FileSink output in multi files
When an execute job in batch mode in flink the fileSink generates multiple files by the parallel number but I want only the output in one file without changing the parallel number
How can I do that ?
...