You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: dss-orchestrator/orchestrators/dss-workflow/dss-workflow-server/src/main/java/com/webank/wedatasphere/dss/workflow/service/impl/DSSFlowServiceImpl.java
+5-5Lines changed: 5 additions & 5 deletions
Original file line number
Diff line number
Diff line change
@@ -1850,23 +1850,23 @@ public List<DataDevelopNodeInfo> dataDevelopNodeResultFilter(DataDevelopNodeRequ
1850
1850
1851
1851
// 新增 Spark 相关属性筛选
1852
1852
if (!StringUtils.isBlank(request.getSparkDriverMemory()) && flag) {
1853
-
flag = request.getSparkDriverMemory().equals(dataDevelopNodeInfo.getSparkDriverMemory());
1853
+
flag = request.getSparkDriverMemory().contains(dataDevelopNodeInfo.getSparkDriverMemory());
1854
1854
}
1855
1855
1856
1856
if (!StringUtils.isBlank(request.getSparkExecutorMemory()) && flag) {
1857
-
flag = request.getSparkExecutorMemory().equals(dataDevelopNodeInfo.getSparkExecutorMemory());
1857
+
flag = request.getSparkExecutorMemory().contains(dataDevelopNodeInfo.getSparkExecutorMemory());
1858
1858
}
1859
1859
1860
1860
if (!StringUtils.isBlank(request.getSparkExecutorCore()) && flag) {
1861
-
flag = request.getSparkExecutorCore().equals(dataDevelopNodeInfo.getSparkExecutorCore());
1861
+
flag = request.getSparkExecutorCore().contains(dataDevelopNodeInfo.getSparkExecutorCore());
1862
1862
}
1863
1863
1864
1864
if (!StringUtils.isBlank(request.getSparkConf()) && flag) {
1865
-
flag = request.getSparkConf().equals(dataDevelopNodeInfo.getSparkConf());
1865
+
flag = request.getSparkConf().contains(dataDevelopNodeInfo.getSparkConf());
1866
1866
}
1867
1867
1868
1868
if (!StringUtils.isBlank(request.getSparkExecutorInstances()) && flag) {
1869
-
flag = request.getSparkExecutorInstances().equals(dataDevelopNodeInfo.getSparkExecutorInstances());
1869
+
flag = request.getSparkExecutorInstances().contains(dataDevelopNodeInfo.getSparkExecutorInstances());
0 commit comments