Apache DolphinScheduler-1.3.9源码分析(二)

server/2024/10/16 2:13:43/

引言

随着大数据的发展,任务调度系统成为了数据处理和管理中至关重要的部分。Apache DolphinScheduler 是一款优秀的开源分布式工作流调度平台,在大数据场景中得到广泛应用。

在本文中,我们将对 Apache DolphinScheduler 1.3.9 版本的源码进行深入分析,主要分析一下Master和Worker的交互设计。

感兴趣的朋友也可以回顾我们上一篇文章:Apache DolphinScheduler-1.3.9源码分析(一)

Worker配置文件

# worker listener port
worker.listen.port=1234# worker execute thread number to limit task instances in parallel
# worker可并行的任务数限制
worker.exec.threads=100# worker heartbeat interval, the unit is second
# worker发送心跳间隔
worker.heartbeat.interval=10# worker max cpuload avg, only higher than the system cpu load average, worker server can be dispatched tasks. default value -1: the number of cpu cores * 2
# worker最大cpu平均负载,只有系统cpu平均负载低于该值,才能执行任务
# 默认值为-1,则最大cpu平均负载=系统cpu核数 * 2
worker.max.cpuload.avg=-1# worker reserved memory, only lower than system available memory, worker server can be dispatched tasks. default value 0.3, the unit is G
# worker的预留内存,只有当系统可用内存大于等于该值,才能执行任务,单位为GB
# 默认0.3G
worker.reserved.memory=0.3# default worker groups separated by comma, like 'worker.groups=default,test'
# 工作组名称,多个用,隔开
worker.groups=default

WorkerServer启动

public void run() {// init remoting serverNettyServerConfig serverConfig = new NettyServerConfig();serverConfig.setListenPort(workerConfig.getListenPort());this.nettyRemotingServer = new NettyRemotingServer(serverConfig);this.nettyRemotingServer.registerProcessor(CommandType.TASK_EXECUTE_REQUEST, new TaskExecuteProcessor());this.nettyRemotingServer.registerProcessor(CommandType.TASK_KILL_REQUEST, new TaskKillProcessor());this.nettyRemotingServer.registerProcessor(CommandType.DB_TASK_ACK, new DBTaskAckProcessor());this.nettyRemotingServer.registerProcessor(CommandType.DB_TASK_RESPONSE, new DBTaskResponseProcessor());this.nettyRemotingServer.start();// worker registrytry {this.workerRegistry.registry();this.workerRegistry.getZookeeperRegistryCenter().setStoppable(this);Set<String> workerZkPaths = this.workerRegistry.getWorkerZkPaths();this.workerRegistry.getZookeeperRegistryCenter().getRegisterOperator().handleDeadServer(workerZkPaths, ZKNodeType.WORKER, Constants.DELETE_ZK_OP);} catch (Exception e) {logger.error(e.getMessage(), e);throw new RuntimeException(e);}// retry report task statusthis.retryReportTaskStatusThread.start();/*** register hooks, which are called before the process exits*/Runtime.getRuntime().addShutdownHook(new Thread(() -> {if (Stopper.isRunning()) {close("shutdownHook");}}));
}
注册四个Command:
  1. TASK_EXECUTE_REQUEST:task执行请求
  2. TASK_KILL_REQUEST:task停止请求
  3. DB_TASK_ACK:Worker接受到Master的调度请求,回应master
  4. DB_TASK_RESPONSE:
  • 注册WorkerServer到Zookeeper,并发送心跳
  • 报告Task执行状态

RetryReportTaskStatusThread

这是一个兜底机制,主要负责定时轮询向Master汇报任务的状态,直到Master回复状态的ACK,避免任务状态丢失;

每隔5分钟,检查一下responceCache中的ACK Cache和Response Cache是否为空,如果不为空则向Master发送ack_commandresponse command请求。

public void run() {ResponceCache responceCache = ResponceCache.get();while (Stopper.isRunning()){// sleep 5 minutesThreadUtils.sleep(RETRY_REPORT_TASK_STATUS_INTERVAL);try {if (!responceCache.getAckCache().isEmpty()){Map<Integer,Command> ackCache =  responceCache.getAckCache();for (Map.Entry<Integer, Command> entry : ackCache.entrySet()){Integer taskInstanceId = entry.getKey();Command ackCommand = entry.getValue();taskCallbackService.sendAck(taskInstanceId,ackCommand);}}if (!responceCache.getResponseCache().isEmpty()){Map<Integer,Command> responseCache =  responceCache.getResponseCache();for (Map.Entry<Integer, Command> entry : responseCache.entrySet()){Integer taskInstanceId = entry.getKey();Command responseCommand = entry.getValue();taskCallbackService.sendResult(taskInstanceId,responseCommand);}}}catch (Exception e){logger.warn("retry report task status error", e);}}
}

Master与Worker的交互设计

Apache DolphinScheduler Master和Worker模块是两个独立的JVM进程,可以部署在不同的服务器上,Master与Worker的通信都是通过Netty实现RPC交互的,一共用到7种处理器。

模块处理器作用
mastermasterTaskResponseProcessor处理TaskExecuteResponseCommand消息,将消息添加到TaskResponseService的任务响应队列中
mastermasterTaskAckProcessor处理TaskExecuteAckCommand消息,将消息添加到TaskResponseService的任务响应队列中
mastermasterTaskKillResponseProcessor处理TaskKillResponseCommand消息,并在日志中打印消息内容
workerworkerTaskExecuteProcessor处理TaskExecuteRequestCommand消息,并发送TaskExecuteAckCommand到master,提交任务执行
workerworkerTaskKillProcessor处理TaskKillRequestCommand消息,调用kill -9 pid杀死任务对应的进程,并向master发送TaskKillResponseCommand消息
workerworkerDBTaskAckProcessor处理DBTaskAckCommand消息,针对执行成功的任务,从ResponseCache中删除
workerworkerDBTaskResponseProcessor处理DBTaskResponseCommand消息,针对执行成功的任务,从ResponseCache中删除

分发任务如何交互

master#TaskPriorityQueueConsumer

Master任务里有一个TaskPriorityQueueConsumer,会从TaskPriorityQueue里每次取3个Task分发给Worker执行,这里会创建TaskExecuteRequestCommand

TaskPriorityQueueConsumer#run()

@Override
public void run() {List<TaskPriority> failedDispatchTasks = new ArrayList<>();while (Stopper.isRunning()){try {// 每一批次分发任务数量,master.dispatch.task.num = 3int fetchTaskNum = masterConfig.getMasterDispatchTaskNumber();failedDispatchTasks.clear();for(int i = 0; i < fetchTaskNum; i++){if(taskPriorityQueue.size() <= 0){Thread.sleep(Constants.SLEEP_TIME_MILLIS);continue;}// if not task , blocking here// 从队列里面获取taskTaskPriority taskPriority = taskPriorityQueue.take();// 分发给worker执行boolean dispatchResult = dispatch(taskPriority);if(!dispatchResult){failedDispatchTasks.add(taskPriority);}}if (!failedDispatchTasks.isEmpty()) {// 分发失败的任务,需要重新加入队列中,等待重新分发for (TaskPriority dispatchFailedTask : failedDispatchTasks) {taskPriorityQueue.put(dispatchFailedTask);}// If there are tasks in a cycle that cannot find the worker group,// sleep for 1 secondif (taskPriorityQueue.size() <= failedDispatchTasks.size()) {TimeUnit.MILLISECONDS.sleep(Constants.SLEEP_TIME_MILLIS);}}}catch (Exception e){logger.error("dispatcher task error",e);}}
}

dispatcher

/*** dispatch task** @param taskPriority taskPriority* @return result*/
protected boolean dispatch(TaskPriority taskPriority) {boolean result = false;try {int taskInstanceId = taskPriority.getTaskId();TaskExecutionContext context = getTaskExecutionContext(taskInstanceId);// 这里创建TaskExecuteRequestCommandExecutionContext executionContext = new ExecutionContext(context.toCommand(), ExecutorType.WORKER, context.getWorkerGroup());if (taskInstanceIsFinalState(taskInstanceId)){// when task finish, ignore this task, there is no need to dispatch anymorereturn true;}else{// 分发任务// 分发算法支持:低负载优先算法,随机算法, 轮询算法。result = dispatcher.dispatch(executionContext);}} catch (ExecuteException e) {logger.error("dispatch error: {}",e.getMessage());}return result;
}

TaskExecutionContext

// 摘录自org.apache.dolphinscheduler.server.entity.TaskExecutionContext#toCommand
public Command toCommand(){TaskExecuteRequestCommand requestCommand = new TaskExecuteRequestCommand();requestCommand.setTaskExecutionContext(FastJsonSerializer.serializeToString(this));return requestCommand.convert2Command();
}

分发算法实现

随机算法

public class RandomSelector<T> implements Selector<T> {private final Random random = new Random();public T select(final Collection<T> source) {if (source == null || source.size() == 0) {throw new IllegalArgumentException("Empty source.");}if (source.size() == 1) {return (T) source.toArray()[0];}int size = source.size();int randomIndex = random.nextInt(size);return (T) source.toArray()[randomIndex];}}

轮询算法

public class RoundRobinSelector<T> implements Selector<T> {private final AtomicInteger index = new AtomicInteger(0);public T select(Collection<T> source) {if (source == null || source.size() == 0) {throw new IllegalArgumentException("Empty source.");}if (source.size() == 1) {return (T)source.toArray()[0];}int size = source.size();return (T) source.toArray()[index.getAndIncrement() % size];}
}

低负载优先算法

public class LowerWeightRoundRobin implements Selector<HostWeight>{public HostWeight select(Collection<HostWeight> sources){int totalWeight = 0;int lowWeight = 0;HostWeight lowerNode = null;for (HostWeight hostWeight : sources) {totalWeight += hostWeight.getWeight();hostWeight.setCurrentWeight(hostWeight.getCurrentWeight() + hostWeight.getWeight());if (lowerNode == null || lowWeight > hostWeight.getCurrentWeight() ) {lowerNode = hostWeight;lowWeight = hostWeight.getCurrentWeight();}}lowerNode.setCurrentWeight(lowerNode.getCurrentWeight() + totalWeight);return lowerNode;}
}

TaskExecuteRequestCommand

TaskExecuteProcessor
构造方法
public TaskExecuteProcessor() {this.taskCallbackService = SpringApplicationContext.getBean(TaskCallbackService.class);this.workerConfig = SpringApplicationContext.getBean(WorkerConfig.class);// worker.exec.threads,默认100this.workerExecService = ThreadUtils.newDaemonFixedThreadExecutor("Worker-Execute-Thread", workerConfig.getWorkerExecThreads());this.taskExecutionContextCacheManager = SpringApplicationContext.getBean(TaskExecutionContextCacheManagerImpl.class);
}
process()方法
public void process(Channel channel, Command command) {Preconditions.checkArgument(CommandType.TASK_EXECUTE_REQUEST == command.getType(),String.format("invalid command type : %s", command.getType()));// 序列化TaskExecuteRequestCommandTaskExecuteRequestCommand taskRequestCommand = FastJsonSerializer.deserialize(command.getBody(), TaskExecuteRequestCommand.class);logger.info("received command : {}", taskRequestCommand);if (taskRequestCommand == null) {logger.error("task execute request command is null");return;}String contextJson = taskRequestCommand.getTaskExecutionContext();TaskExecutionContext taskExecutionContext = JSONObject.parseObject(contextJson, TaskExecutionContext.class);if (taskExecutionContext == null) {logger.error("task execution context is null");return;}// 存入taskExecutionContextCacheManagersetTaskCache(taskExecutionContext);// 创建任务日志Logger taskLogger = LoggerFactory.getLogger(LoggerUtils.buildTaskId(LoggerUtils.TASK_LOGGER_INFO_PREFIX,taskExecutionContext.getProcessDefineId(),taskExecutionContext.getProcessInstanceId(),taskExecutionContext.getTaskInstanceId()));taskExecutionContext.setHost(NetUtils.getAddr(workerConfig.getListenPort()));taskExecutionContext.setStartTime(new Date());taskExecutionContext.setLogPath(getTaskLogPath(taskExecutionContext));// local execute pathString execLocalPath = getExecLocalPath(taskExecutionContext);logger.info("task instance local execute path : {}", execLocalPath);taskExecutionContext.setExecutePath(execLocalPath);// ThreadLocal存储任务日志FileUtils.taskLoggerThreadLocal.set(taskLogger);try {// 创建执行FileUtils.createWorkDirAndUserIfAbsent(execLocalPath, taskExecutionContext.getTenantCode());} catch (Throwable ex) {String errorLog = String.format("create execLocalPath : %s", execLocalPath);LoggerUtils.logError(Optional.ofNullable(logger), errorLog, ex);LoggerUtils.logError(Optional.ofNullable(taskLogger), errorLog, ex);taskExecutionContextCacheManager.removeByTaskInstanceId(taskExecutionContext.getTaskInstanceId());}FileUtils.taskLoggerThreadLocal.remove();taskCallbackService.addRemoteChannel(taskExecutionContext.getTaskInstanceId(),new NettyRemoteChannel(channel, command.getOpaque()));// 向master发送TaskExecuteAckCommandthis.doAck(taskExecutionContext);// submit taskworkerExecService.submit(new TaskExecuteThread(taskExecutionContext, taskCallbackService, taskLogger));
}private void doAck(TaskExecutionContext taskExecutionContext){// tell master that task is in executingTaskExecuteAckCommand ackCommand = buildAckCommand(taskExecutionContext);ResponceCache.get().cache(taskExecutionContext.getTaskInstanceId(),ackCommand.convert2Command(),Event.ACK);taskCallbackService.sendAck(taskExecutionContext.getTaskInstanceId(), ackCommand.convert2Command());
}

TaskExecuteThread

构造方法
public TaskExecuteThread(TaskExecutionContext taskExecutionContext, TaskCallbackService taskCallbackService, Logger taskLogger) {this.taskExecutionContext = taskExecutionContext;this.taskCallbackService = taskCallbackService;this.taskExecutionContextCacheManager = SpringApplicationContext.getBean(TaskExecutionContextCacheManagerImpl.class);this.taskLogger = taskLogger;
}
运行方法
public void run() {TaskExecuteResponseCommand responseCommand = new TaskExecuteResponseCommand(taskExecutionContext.getTaskInstanceId());try {logger.info("script path : {}", taskExecutionContext.getExecutePath());// task nodeTaskNode taskNode = JSONObject.parseObject(taskExecutionContext.getTaskJson(), TaskNode.class);// copy hdfs/minio file to local// 下载需要的资源,例如Spark/Flink jar,udf等downloadResource(taskExecutionContext.getExecutePath(),taskExecutionContext.getResources(),logger);taskExecutionContext.setTaskParams(taskNode.getParams());taskExecutionContext.setEnvFile(CommonUtils.getSystemEnvPath());taskExecutionContext.setDefinedParams(getGlobalParamsMap());// set task timeoutsetTaskTimeout(taskExecutionContext, taskNode);taskExecutionContext.setTaskAppId(String.format("%s_%s_%s",taskExecutionContext.getProcessDefineId(),taskExecutionContext.getProcessInstanceId(),taskExecutionContext.getTaskInstanceId()));// 创建任务task = TaskManager.newTask(taskExecutionContext, taskLogger);// 初始化任务task.init();// 构建任务所需要的参数preBuildBusinessParams();// 执行任务task.handle();// 任务执行完成后的动作task.after();responseCommand.setStatus(task.getExitStatus().getCode());responseCommand.setEndTime(new Date());responseCommand.setProcessId(task.getProcessId());responseCommand.setAppIds(task.getAppIds());logger.info("task instance id : {},task final status : {}", taskExecutionContext.getTaskInstanceId(), task.getExitStatus());} catch (Exception e) {logger.error("task scheduler failure", e);// 如果出现异常,kill taskkill();responseCommand.setStatus(ExecutionStatus.FAILURE.getCode());responseCommand.setEndTime(new Date());responseCommand.setProcessId(task.getProcessId());responseCommand.setAppIds(task.getAppIds());} finally {// 从cache中去除任务执行上下文。taskExecutionContextCacheManager.removeByTaskInstanceId(taskExecutionContext.getTaskInstanceId());// 缓存responseCommandResponceCache.get().cache(taskExecutionContext.getTaskInstanceId(), responseCommand.convert2Command(), Event.RESULT);// 向master发送ResponseCommandtaskCallbackService.sendResult(taskExecutionContext.getTaskInstanceId(), responseCommand.convert2Command());// 清除task执行路径clearTaskExecPath();}
}

master#TaskResponseService

Worker在正常执行分发任务的时候,会向Master发送ACK Command 和 Response Command。

在Master中,则由TaskAckProcessorTaskResponseProcessor进行处理。

TaskAckProcessor

public void process(Channel channel, Command command) {Preconditions.checkArgument(CommandType.TASK_EXECUTE_ACK == command.getType(), String.format("invalid command type : %s", command.getType()));TaskExecuteAckCommand taskAckCommand = FastJsonSerializer.deserialize(command.getBody(), TaskExecuteAckCommand.class);logger.info("taskAckCommand : {}", taskAckCommand);// 添加缓存taskInstanceCacheManager.cacheTaskInstance(taskAckCommand);String workerAddress = ChannelUtils.toAddress(channel).getAddress();ExecutionStatus ackStatus = ExecutionStatus.of(taskAckCommand.getStatus());// TaskResponseEventTaskResponseEvent taskResponseEvent = TaskResponseEvent.newAck(ackStatus,taskAckCommand.getStartTime(),workerAddress,taskAckCommand.getExecutePath(),taskAckCommand.getLogPath(),taskAckCommand.getTaskInstanceId(),channel);// 主要处理逻辑taskResponseService.addResponse(taskResponseEvent);
}

TaskResponseProcessor

public void process(Channel channel, Command command) {Preconditions.checkArgument(CommandType.TASK_EXECUTE_RESPONSE == command.getType(), String.format("invalid command type : %s", command.getType()));TaskExecuteResponseCommand responseCommand = FastJsonSerializer.deserialize(command.getBody(), TaskExecuteResponseCommand.class);logger.info("received command : {}", responseCommand);// 缓存taskInstanceCacheManager.cacheTaskInstance(responseCommand);// TaskResponseEventTaskResponseEvent taskResponseEvent = TaskResponseEvent.newResult(ExecutionStatus.of(responseCommand.getStatus()),responseCommand.getEndTime(),responseCommand.getProcessId(),responseCommand.getAppIds(),responseCommand.getTaskInstanceId(),channel);// 主要处理逻辑taskResponseService.addResponse(taskResponseEvent);
}

TaskResponseService

通过TaskResponseProcessorTaskAckProcessor发现,其主要逻辑都在TaskResponseService类中,而TaskResponseService中处理事件,是通过TaskResponseWorker线程实现的。

// TaskResponseEvent队列是阻塞队列
private final BlockingQueue<TaskResponseEvent> eventQueue = new LinkedBlockingQueue<>(5000);class TaskResponseWorker extends Thread {@Overridepublic void run() {while (Stopper.isRunning()){try {// 如果没有任务事件,则会阻塞在这里TaskResponseEvent taskResponseEvent = eventQueue.take();// 任务实例状态持久化到数据库persist(taskResponseEvent);} catch (InterruptedException e){break;} catch (Exception e){logger.error("persist task error",e);}}logger.info("TaskResponseWorker stopped");}}/*** persist  taskResponseEvent* @param taskResponseEvent taskResponseEvent*/private void persist(TaskResponseEvent taskResponseEvent){Event event = taskResponseEvent.getEvent();Channel channel = taskResponseEvent.getChannel();switch (event){case ACK:try {TaskInstance taskInstance = processService.findTaskInstanceById(taskResponseEvent.getTaskInstanceId());if (taskInstance != null) {ExecutionStatus status = taskInstance.getState().typeIsFinished() ? taskInstance.getState() : taskResponseEvent.getState();processService.changeTaskState(status,taskResponseEvent.getStartTime(),taskResponseEvent.getWorkerAddress(),taskResponseEvent.getExecutePath(),taskResponseEvent.getLogPath(),taskResponseEvent.getTaskInstanceId());}// 向worker发送DB_TASK_ACK请求DBTaskAckCommand taskAckCommand = new DBTaskAckCommand(ExecutionStatus.SUCCESS.getCode(), taskResponseEvent.getTaskInstanceId());channel.writeAndFlush(taskAckCommand.convert2Command());}catch (Exception e){logger.error("worker ack master error",e);DBTaskAckCommand taskAckCommand = new DBTaskAckCommand(ExecutionStatus.FAILURE.getCode(),-1);channel.writeAndFlush(taskAckCommand.convert2Command());}break;case RESULT:try {TaskInstance taskInstance = processService.findTaskInstanceById(taskResponseEvent.getTaskInstanceId());if (taskInstance != null){processService.changeTaskState(taskResponseEvent.getState(),taskResponseEvent.getEndTime(),taskResponseEvent.getProcessId(),taskResponseEvent.getAppIds(),taskResponseEvent.getTaskInstanceId());}// 向worker发送DB_TASK_RESPONSE请求DBTaskResponseCommand taskResponseCommand = new DBTaskResponseCommand(ExecutionStatus.SUCCESS.getCode(),taskResponseEvent.getTaskInstanceId());channel.writeAndFlush(taskResponseCommand.convert2Command());}catch (Exception e){logger.error("worker response master error",e);DBTaskResponseCommand taskResponseCommand = new DBTaskResponseCommand(ExecutionStatus.FAILURE.getCode(),-1);channel.writeAndFlush(taskResponseCommand.convert2Command());}break;default:throw new IllegalArgumentException("invalid event type : " + event);}}

Worker#DBTaskAckProcessor和DBTaskResponseProcessor

Worker接受到Master的db_task_ack commanddb_task_response command,对应的处理器为DBTaskAckProcessorDBTaskResponseProcessor,其逻辑都是从ResponceCache删除对应的task instance command

DBTaskAckProcessor

public void process(Channel channel, Command command) {Preconditions.checkArgument(CommandType.DB_TASK_ACK == command.getType(),String.format("invalid command type : %s", command.getType()));DBTaskAckCommand taskAckCommand = FastJsonSerializer.deserialize(command.getBody(), DBTaskAckCommand.class);if (taskAckCommand == null){return;}if (taskAckCommand.getStatus() == ExecutionStatus.SUCCESS.getCode()){ResponceCache.get().removeAckCache(taskAckCommand.getTaskInstanceId());}
}

DBTaskResponseProcessor

public void process(Channel channel, Command command) {Preconditions.checkArgument(CommandType.DB_TASK_RESPONSE == command.getType(),String.format("invalid command type : %s", command.getType()));DBTaskResponseCommand taskResponseCommand = FastJsonSerializer.deserialize(command.getBody(), DBTaskResponseCommand.class);if (taskResponseCommand == null){return;}if (taskResponseCommand.getStatus() == ExecutionStatus.SUCCESS.getCode()){ResponceCache.get().removeResponseCache(taskResponseCommand.getTaskInstanceId());}
}

停止任务如何交互

MasterTaskExecThread#waitTaskQuit

public Boolean waitTaskQuit(){// query new statetaskInstance = processService.findTaskInstanceById(taskInstance.getId());while (Stopper.isRunning()){try {// 省略代码...// task instance add queue , waiting worker to kill// 如果master接受到cancal请求,或者工作流状态为准备停止的状态// master会给worker发送kill request command请求if(this.cancel || this.processInstance.getState() == ExecutionStatus.READY_STOP){cancelTaskInstance();}// 省略代码...} catch (Exception e) {// 省略代码...}}return true;
}private void cancelTaskInstance() throws Exception{if(alreadyKilled){return;}alreadyKilled = true;taskInstance = processService.findTaskInstanceById(taskInstance.getId());if(StringUtils.isBlank(taskInstance.getHost())){taskInstance.setState(ExecutionStatus.KILL);taskInstance.setEndTime(new Date());processService.updateTaskInstance(taskInstance);return;}// 构造TaskKillRequestCommandTaskKillRequestCommand killCommand = new TaskKillRequestCommand();killCommand.setTaskInstanceId(taskInstance.getId());ExecutionContext executionContext = new ExecutionContext(killCommand.convert2Command(), ExecutorType.WORKER);Host host = Host.of(taskInstance.getHost());executionContext.setHost(host);nettyExecutorManager.executeDirectly(executionContext);logger.info("master kill taskInstance name :{} taskInstance id:{}",taskInstance.getName(), taskInstance.getId() );
}

Worker#TaskKillProcessor

TaskKillProcessor用于处理Master发送的Kill request command

public void process(Channel channel, Command command) {Preconditions.checkArgument(CommandType.TASK_KILL_REQUEST == command.getType(), String.format("invalid command type : %s", command.getType()));TaskKillRequestCommand killCommand = FastJsonSerializer.deserialize(command.getBody(), TaskKillRequestCommand.class);logger.info("received kill command : {}", killCommand);Pair<Boolean, List<String>> result = doKill(killCommand);taskCallbackService.addRemoteChannel(killCommand.getTaskInstanceId(),new NettyRemoteChannel(channel, command.getOpaque()));// 向master发送kill response commandTaskKillResponseCommand taskKillResponseCommand = buildKillTaskResponseCommand(killCommand,result);taskCallbackService.sendResult(taskKillResponseCommand.getTaskInstanceId(), taskKillResponseCommand.convert2Command());taskExecutionContextCacheManager.removeByTaskInstanceId(taskKillResponseCommand.getTaskInstanceId());
}private Pair<Boolean, List<String>> doKill(TaskKillRequestCommand killCommand){boolean processFlag = true;List<String> appIds = Collections.emptyList();int taskInstanceId = killCommand.getTaskInstanceId();TaskExecutionContext taskExecutionContext = taskExecutionContextCacheManager.getByTaskInstanceId(taskInstanceId);try {Integer processId = taskExecutionContext.getProcessId();if (processId.equals(0)) {taskExecutionContextCacheManager.removeByTaskInstanceId(taskInstanceId);logger.info("the task has not been executed and has been cancelled, task id:{}", taskInstanceId);return Pair.of(true, appIds);}// 执行Kill -9 命令直接删除进程// spark or flink如果是提交到集群,暂时Kill不掉String pidsStr = ProcessUtils.getPidsStr(taskExecutionContext.getProcessId());if (StringUtils.isNotEmpty(pidsStr)) {String cmd = String.format("sudo kill -9 %s", ProcessUtils.getPidsStr(taskExecutionContext.getProcessId()));logger.info("process id:{}, cmd:{}", taskExecutionContext.getProcessId(), cmd);OSUtils.exeCmd(cmd);}} catch (Exception e) {processFlag = false;logger.error("kill task error", e);}// find log and kill yarn jobPair<Boolean, List<String>> yarnResult = killYarnJob(Host.of(taskExecutionContext.getHost()).getIp(),taskExecutionContext.getLogPath(),taskExecutionContext.getExecutePath(),taskExecutionContext.getTenantCode());return Pair.of(processFlag && yarnResult.getLeft(), yarnResult.getRight());
}

master#TaskKillResponseProcessor

TaskKillResponseProcessor用于master处理worker停止任务的响应请求。

public void process(Channel channel, Command command) {Preconditions.checkArgument(CommandType.TASK_KILL_RESPONSE == command.getType(), String.format("invalid command type : %s", command.getType()));TaskKillResponseCommand responseCommand = FastJsonSerializer.deserialize(command.getBody(), TaskKillResponseCommand.class);logger.info("received task kill response command : {}", responseCommand);
}

通过对 Apache DolphinScheduler 1.3.9 的源码分析,我们深入了解了其核心模块的设计和实现。

如果你对 Apache DolphinScheduler 的源码有兴趣,可以深入研究其任务调度策略的细节部分,或者根据自身业务场景进行二次开发,充分发挥 DolphinScheduler 的调度能力。

本文完!

本文由 白鲸开源科技 提供发布支持!


http://www.ppmy.cn/server/132489.html

相关文章

凡事预则立,不预则废

在交易的竞技场上&#xff0c;每位交易员都拥有自己的一套打法&#xff0c;这些独特的交易风格不仅塑造了他们的个人特点&#xff0c;更是他们成功的关键所在。今天采访的Eagle Trader交易员刘先生&#xff0c;就给我一种很稳妥的感觉&#xff0c;那么&#xff0c;刘先生的“稳…

AWS注册时常见错误处理

引言 创建AWS账号是使用AWS云服务的第一步&#xff0c;但在注册过程中可能会遇到一些常见的问题。本文中九河云将帮助您排查和解决在创建AWS账户时可能遇到的一些常见问题&#xff0c;包括未接到验证电话、最大失败尝试次数错误以及账户激活延迟等。 常见问题及解决方法 1. …

【开发工具】K8s经验: 在k8s中 服务1部署到了k8s,那服务2也在k8s中的某个dockers中,服务1和服务2能互相找到吗 ?

今天在开发中遇到了需要调用其他dubbo服务中的一个接口&#xff0c;首先通过注解DubboReference来引入另外一个dubbo服务&#xff0c;但是还需要设置url吗&#xff1f; 下面是思考经验&#xff1a;希望能帮到大家 服务1已经被部署到了Kubernetes&#xff08;k8s&#xff09;集…

7.跨学科应用

1. 跨学科应用 人工智能药物研发是指利用人工智能技术研发新药物的过程和方法&#xff0c;人工智能拔术可以用于药物研发的不同阶段&#xff0c;包括药物发现、分子设计、药效预测、毒性评估等可以加速药物研发的过程、提高药物的效力和安全性。下面介绍几种常见的与人…

sublime配置(竞赛向)

我也想要有jiangly一样的sublime 先决条件 首先&#xff0c;到官网上下载最新的sublime4&#xff0c;然后在mingw官网上下载最新的mingw64 mingw64官网&#xff1a;左边菜单栏点击dowloads,然后选择MinGW-W64-builds(可能会有点慢)——然后有时候会变成选LLVM-minGW,接着选择…

2024VDC蓝河分会场:蓝河操作系统2 全栈自研 为AI而生

10月10日&#xff0c; 以“同心同行”为主题的2024vivo开发者大会在深圳成功举办&#xff0c;在同期举办的蓝河分会场上&#xff0c;vivo多位专家及产业界、学术界伙伴分享了在AGI时代下&#xff0c;蓝河操作系统带来的技术创新与实践&#xff0c;vivo希望携各方共建生态&#…

UIP协议栈 TCP Server Client通信成功案例

文章目录 这里边有相当好的 [UIP 文档资料&#xff0c;文档位置在仓库的UIP/uip doc &#xff0c;括号内是仓库地址&#xff08;https://gitee.com/free-people-in-time-and-space/net-work-learn-note.git &#xff09;TCP Server1.main循环里做的事2.以下是main循环里相关函数…

JQuery基本介绍和使用方法

JQuery基本介绍和使用方法 W3C 标准给我们提供了⼀系列的函数, 让我们可以操作: ⽹⻚内容⽹⻚结构⽹⻚样式 但是原⽣的JavaScript提供的API操作DOM元素时, 代码⽐较繁琐, 冗⻓. 我们可以使⽤JQuery来操作⻚⾯对象. jQuery是⼀个快速、简洁且功能丰富的JavaScript框架, 于20…