为了更快的定位您的问题,请提供以下信息,谢谢
【详述】
一个BE节点忽然挂掉,通过查询be.out日志发现如下信息:
Caused by: org.apache.hadoop.ipc.RemoteException(java.io.FileNotFoundException): File does not exist: /user/hive/warehouse/ods_sap.db/vbap/part_dt=INIT/000012_0
at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:85)
at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:75)
at org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getBlockLocations(FSDirStatAndListingOp.java:152)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1909)
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:735)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:415)
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:869)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:815)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2675)
at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584)
at org.apache.hadoop.ipc.Client.call(Client.java:1529)
at org.apache.hadoop.ipc.Client.call(Client.java:1426)
at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258)
at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139)
at com.sun.proxy.$Proxy11.getBlockLocations(Unknown Source)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$getBlockLocations$0(ClientNamenodeProtocolTranslatorPB.java:340)
at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getBlockLocations(ClientNamenodeProtocolTranslatorPB.java:340)
at jdk.internal.reflect.GeneratedMethodAccessor15.invoke(Unknown Source)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437)
at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170)
at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162)
at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366)
at com.sun.proxy.$Proxy12.getBlockLocations(Unknown Source)
at org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:931)
... 7 more
3.3.9 RELEASE (build dfae8f9)
query_id:caee0940-e5ee-11f0-9de5-005056a4879f, fragment_instance:caee0940-e5ee-11f0-9de5-005056a487c8
Hive file path: 000007_0, partition id: 8, length: 244469282, offset: 0
tracker:process consumption: 98600360080
tracker:jemalloc_metadata consumption: 4163733712
tracker:query_pool consumption: 311684736
tracker:query_pool/connector_scan consumption: 2254609284
tracker:load consumption: 0
tracker:metadata consumption: 3882793849
tracker:tablet_metadata consumption: 16945915
tracker:rowset_metadata consumption: 24949410
tracker:segment_metadata consumption: 265567318
tracker:column_metadata consumption: 3575331206
tracker:tablet_schema consumption: 188995
tracker:segment_zonemap consumption: 39660228
tracker:short_key_index consumption: 221973942
tracker:column_zonemap_index consumption: 1122772934
tracker:ordinal_index consumption: 2181466208
tracker:bitmap_index consumption: 0
tracker:bloom_filter_index consumption: 0
tracker:compaction consumption: 996712
tracker:schema_change consumption: 0
tracker:column_pool consumption: 0
tracker:page_cache consumption: 43726916048
tracker:jit_cache consumption: 4616
tracker:update consumption: 34344122432
tracker:chunk_allocator consumption: 0
tracker:passthrough consumption: 0
tracker:clone consumption: 0
tracker:consistency consumption: 0
tracker:datacache consumption: 0
tracker:replication consumption: 0
*** Aborted at 1767147423 (unix time) try “date -d @1767147423” if you are using GNU date ***
PC: @ 0x70984ea starrocks::parquet::StoredColumnReaderImpl::get_dict_values(starrocks::Column*)
*** SIGSEGV (@0x0) received by PID 47263 (TID 0x7ff7376d1700) from PID 0; stack trace: ***
@ 0x7ff903eb620b __pthread_once_slow
@ 0x7bed140 google::(anonymous namespace)::FailureSignalHandler(int, siginfo_t*, void*)
@ 0x7ff904d01e2b os::Linux::chained_handler(int, siginfo*, void*)
@ 0x7ff904d0699e JVM_handle_linux_signal
@ 0x7ff904cfac08 signalHandler(int, siginfo*, void*)
@ 0x7ff903ebf630 (/usr/lib64/libpthread-2.17.so+0xf62f)
@ 0x70984ea starrocks::parquet::StoredColumnReaderImpl::get_dict_values(starrocks::Column*)
@ 0x708836a starrocks::parquet::ColumnDictFilterContext::rewrite_conjunct_ctxs_to_predicate(starrocks::parquet::StoredColumnReader*, bool*)
@ 0x708f78d starrocks::parquet::ScalarColumnReader::rewrite_conjunct_ctxs_to_predicate(bool*, std::vector<std::__cxx11::basic_string<char, std::char_traits, std::allocator >, std::allocator<std::__cxx11::basic_string<char, std::char_traits, std::allo
@ 0x7083059 starrocks::parquet::GroupReader::_rewrite_conjunct_ctxs_to_predicates(bool*)
@ 0x7085363 starrocks::parquet::GroupReader::prepare()
@ 0x70570ba starrocks::parquet::FileReader::_init_group_readers()
@ 0x7057616 starrocks::parquet::FileReader::init(starrocks::HdfsScannerContext*)
@ 0x6e8a023 starrocks::HdfsParquetScanner::do_open(starrocks::RuntimeState*)
@ 0x6e7bb30 starrocks::HdfsScanner::open(starrocks::RuntimeState*)
@ 0x6e0a190 starrocks::connector::HiveDataSource::_init_scanner(starrocks::RuntimeState*)
@ 0x6e0ba12 starrocks::connector::HiveDataSource::open(starrocks::RuntimeState*)
@ 0x432bb74 starrocks::pipeline::ConnectorChunkSource::_open_data_source(starrocks::RuntimeState*, bool*)
@ 0x432c2e9 starrocks::pipeline::ConnectorChunkSource::_read_chunk(starrocks::RuntimeState*, std::shared_ptrstarrocks::Chunk)
@ 0x4645ec8 starrocks::pipeline::ChunkSource::buffer_next_batch_chunks_blocking(starrocks::RuntimeState, unsigned long, starrocks::workgroup::WorkGroup const*)
@ 0x431bc4d auto starrocks::pipeline::ScanOperator::_trigger_next_scan(starrocks::RuntimeState*, int)::{lambda(auto:1&)#1}::operator()starrocks::workgroup::YieldContext(starrocks::workgroup::YieldContext&) const [clone .constprop.0]
@ 0x441641b starrocks::workgroup::ScanExecutor::worker_thread()
@ 0x384e543 starrocks::ThreadPool::dispatch_thread()
@ 0x3846fa6 starrocks:
:supervise_thread(void*)
@ 0x7ff903eb7ea5 start_thread
@ 0x7ff9032b896d __clone
通过查询fe.audit找到了对应的SQL语句:
2025-12-31 10:17:03.482+08:00 [query] |Timestamp=1767147421806
|Client=192.168.1.10:63580
|User=bi_search
|AuthorizedUser=‘bi_search’@’%’|
ResourceGroup=default_wg|Catalog=default_catalog|Db=bi_db|State=ERR|ErrorCode=INTERNAL_ERROR|Time=1676|
ScanBytes=0
|ScanRows=0|ReturnRows=0|StmtId=738876
|QueryId=caee0940-e5ee-11f0-9de5-005056a4879f|IsQuery=true|feIp=192.168.1.101
|Stmt=
WITH TAB_MON_LST AS (
SELECT
DATE_FORMAT(DATE_TRUNC(‘month’, DATE_ADD( LAST_DAY( ‘2025-12-31’ ), INTERVAL -5 MONTH)),’%Y%m’) as MON_1,
DATE_FORMAT(DATE_TRUNC(‘month’, DATE_ADD( LAST_DAY( ‘2025-12-31’ ), INTERVAL -4 MONTH)),’%Y%m’) as MON_2,
DATE_FORMAT(DATE_TRUNC(‘month’, DATE_ADD( LAST_DAY( ‘2025-12-31’ ), INTERVAL -3 MONTH)),’%Y%m’) as MON_3,
DATE_FORMAT(DATE_TRUNC(‘month’, DATE_ADD( LAST_DAY( ‘2025-12-31’ ), INTERVAL -2 MONTH)),’%Y%m’) as MON_4,
DATE_FORMAT(DATE_TRUNC(‘month’, DATE_ADD( LAST_DAY( ‘2025-12-31’ ), INTERVAL -1 MONTH)),’%Y%m’) as MON_5,
DATE_FORMAT(DATE_TRUNC(‘month’, DATE_ADD( LAST_DAY( ‘2025-12-31’ ), INTERVAL -0 MONTH)),’%Y%m’) as MON_6 ),
TAB_STOCK_DOC AS (
SELECT G.MBLNR, G.MJAHR, G.ZEILE, F.BUDAT, G.BWART, G.MATNR, G.MENGE, G.AUFNR,
G.WERKS
FROM
DEFAULT_CATALOG.CDC_DB.MKPF F
INNER JOIN DEFAULT_CATALOG.CDC_DB.MSEG G ON F.MBLNR = G.MBLNR AND F.MJAHR = G.MJAHR
WHERE F.BUDAT BETWEEN REPLACE( ‘2025-06-01’ ,’-’,’’) AND REPLACE( ‘2025-12-31’ ,’-’,’’) AND G.WERKS IN (‘1000’, ‘1200’)
AND G.BWART IN (‘261’, ‘262’, ‘951’, ‘952’, ‘961’, ‘962’, ‘971’, ‘972’, ‘201’, ‘202’, ‘231’, ‘232’, ‘101’, ‘102’) AND G.AUFNR <> ‘000000200163’
AND G.MATNR IN ( ‘Q306T2PZ’,‘Q306T2PP’ )
),
TAB_MARC AS (
SELECT
C.MATNR AS MATNR, C.WERKS AS WERKS, C.SOBSL AS SOBSL FROM hive_catalog.ODS_SAP.MARC C WHERE C.WERKS IN (‘1000’, ‘1200’) AND C.MATNR IN
( ‘Q3006T2PZ’,‘Q33006T2PP’ )
),
TAB_DTL AS (
SELECT
P.AUFNR, CASE WHEN V.VBELN IS NOT NULL THEN V.VBELN ELSE VW.VBELN END AS SLS_DOC_NO,
CASE WHEN A.BWART NOT IN (‘251’, ‘252’) AND V.VBELN IS NOT NULL THEN V.MATNR WHEN A.BWART NOT IN (‘251’, ‘252’) AND V.VBELN IS NULL THEN VW.MATNR ELSE ’ ’ END AS SLS_DOC_MAT,
CASE WHEN MC.SOBSL IN (‘82’, ‘83’, ‘84’, ‘85’, ‘86’, ‘94’, ‘96’) AND A.BWART IN (‘261’, ‘262’) THEN A.MENGE
WHEN MC.SOBSL NOT IN (‘82’, ‘83’, ‘84’, ‘85’, ‘86’, ‘94’, ‘96’, ‘90’) AND A.BWART NOT IN (‘101’, ‘102’) THEN A.MENGE ELSE 0 END AS AMT,
A.*,SUBSTR(A.BUDAT,1,6) AS DOC_MON, L.MON_1 AS MON_1,L.MON_2 AS MON_2,L.MON_3 AS MON_3,L.MON_4 AS MON_4,L.MON_5 AS MON_5,L.MON_6 AS MON_6
FROM
TAB_STOCK_DOC A INNER
JOIN TAB_MON_LST L ON 1 = 1 LEFT JOIN DEFAULT_CATALOG.CDC_DB.AFPO P ON P.AUFNR = A.AUFNR
LEFT JOIN hive_catalog.ODS_SAP.VBAP V ON P.KDAUF = V.VBELN AND P.KDPOS = V.POSNR AND P.KDAUF <> ’ ’
LEFT JOIN DEFAULT_CATALOG.CDC_DB.AFPO PW ON PW.AUFNR = P.WEMPF
LEFT JOIN hive_catalog.ODS_SAP.VBAP VW ON PW.KDAUF = VW.VBELN AND PW.KDPOS = ‘000010’ AND PW.KDAUF <> ’ ’
LEFT JOIN TAB_MARC MC ON MC.MATNR = A.MATNR AND MC.WERKS = A.WERKS
)
SELECT
X.MATNR AS MATNR, X.SLS_DOC_MAT AS BUSTY, SUM(CASE WHEN X.MON_1 = X.DOC_MON THEN ROUND(X.AMT,1) ELSE 0 END) AS MENN1, SUM(CASE WHEN X.MON_2 = X.DOC_MON THEN ROUND(X.AMT,1) ELSE 0 END) AS MENN2,
SUM(CASE WHEN X.MON_3 = X.DOC_MON THEN ROUND(X.AMT,1) ELSE 0 END) AS MENN3, SUM(CASE WHEN X.MON_4 = X.DOC_MON THEN ROUND(X.AMT,1) ELSE 0 END) AS MENN4,
SUM(CASE WHEN X.MON_5 = X.DOC_MON THEN ROUND(X.AMT,1) ELSE 0 END) AS MENN5, SUM(CASE WHEN X.MON_6 = X.DOC_MON THEN ROUND(X.AMT,1) ELSE 0 END) AS MENN6, SUM(ROUND(X.AMT,1)) AS MEN16,
ROUND(AVG(Y.AMT),1) AS MENTO, CASE WHEN AVG(Y.AMT) = 0 THEN ‘0%’ ELSE concat(CAST( ROUND(SUM(X.AMT) / AVG(Y.AMT) * 100, 2) AS VARCHAR) ,’%’) END PROPO
FROM
TAB_DTL X
INNER JOIN (SELECT MATNR,SUM(AMT) AS AMT FROM TAB_DTL GROUP BY MATNR) Y ON X.MATNR = Y.MATNR
GROUP BY X.MATNR,X.SLS_DOC_MAT
ORDER BY X.MATNR,X.SLS_DOC_MAT;
|Digest=|PlanCpuCost=1.3097829700500685E10|PlanMemCost=1.943827677329574E9|PendingTimeMs=0|Warehouse=default_warehouse|IsForwardToLeader=false
starrocks版本:3.3.9
【背景】做过哪些操作?
【业务影响】
【是否存算分离】
【StarRocks版本】3.3.9
【集群规模】3fe+3be
【