Cause of accident
Ten million levels of data fall into the partition table, and the inserted partition is set as a dynamic partition
According to the resource usage on yarn, it is found that the memory, CPU and core count of the cluster are normal, but it appears after 88% of the data import task is executed
GC overhead limit exceeded. According to the generated execution log, it is found that there is only map task for data import, and the data volume level is tens of millions
Setting JVM parameters
set mapred.child.java.opts=-Xmx8000m;
set mapreduce.map.java.opts=-Xmx8096m;
set mapreduce.reduce.java.opts=-Xmx8096m;
set mapreduce.map.memory.mb=8096;
set mapreduce.reduce.memory.mb=8096;
By increasing the JVM parameters, you can make the data run past, but as long as the amount of data doubles, an error will still be reported
GC overhead limit exceeded
Optimization scheme:
When inserting data, add cluster by after the insertion condition to the key data fields, disperse the data, and generate a certain amount of reduce tasks to process part of the data
Read More:
- [Solved] Hadoop Mapreduce Error: GC overhead limit exceeded
- [Solved] kettle Error: GC overhead limit exceeded
- IDEA Compile Project Error: GC overhead limit exceeded solution
- Idea Error: java: java.lang.OutOfMemoryError: GC overhead limit exceeded
- [Solved] “HttpSendRequest“ failed, Windows error code=12002 andretry limit (0) exceeded for URL
- Tidb2.1 reports Error statement count 5001 exceeded the transaction limit, autocommit = false
- Impala Find Error: InternalException: Memory limit exceeded: Error occurred on backend
- [Solved] Keil Error: *** ERROR L250: CODE SIZE LIMIT IN RESTRICTED VERSION EXCEEDED
- Hive: Hive partition sorting error [How to Solve]
- Hive Error: FAILED: RuntimeException Error loading hooks(hive.exec.post.hooks): java.lang.ClassNotFoundException: org.apache.atlas.hive.hook.HiveHook
- [Solved] Hive Error: FAILED: Execution Error, return code 3 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
- HIVE Error: Error: Error while processing statement: FAILED: Execution Error, return code 2 from org.apac
- [Solved] Hexo FATAL ERROR: Ineffective mark-compacts near heap limit Allocation failed
- [HBase Error]“java.lang.OutOfMemoryError: Requested array size exceeds VM limit”
- How to Solve DB2 uses Limit Error
- This limit can be set by changing the [search.max_open_scroll_context] setting
- Sqoop exports hive data to MySQL Error [How to Solve]
- [Solved] waterdrop Import hive to clickhouse Error: Too many partitions for single INSERT block (more than 100).
- Etcd Error: mvcc: database space exceeded
- [Solved] Doris BrokerLoad Error: Scan bytes per broker scanner exceed limit: 3221225472