pom:(1.3.0-beta1): 调整合并的代码不兼容问题

This commit is contained in:
opensnail 2024-10-30 22:26:41 +08:00
parent c112d1c9fd
commit 432e35d316
9 changed files with 38 additions and 39 deletions

View File

@ -10,7 +10,6 @@ import com.aizuda.snailjob.client.common.exception.SnailJobClientException;
*/
public abstract class AbstractRequestHandler<R> implements RequestHandler<R> {
protected static final String SHARD_NUM = "shardNum";
/**
* 具体调用
* @return

View File

@ -2,11 +2,11 @@ package com.aizuda.snailjob.client.job.core.handler;
import com.aizuda.snailjob.client.common.rpc.client.RequestBuilder;
import com.aizuda.snailjob.client.job.core.openapi.OpenApiClient;
import com.aizuda.snailjob.common.core.model.NettyResult;
import com.aizuda.snailjob.common.core.model.SnailJobRpcResult;
public interface RequestHandler<R> {
OpenApiClient client = RequestBuilder.<OpenApiClient, NettyResult>newBuilder()
OpenApiClient client = RequestBuilder.<OpenApiClient, SnailJobRpcResult>newBuilder()
.client(OpenApiClient.class)
.async(false)
.build();

View File

@ -1,4 +1,4 @@
package com.aizuda.snailjob.server.job.task.support.handler;
package com.aizuda.snailjob.server.job.task.support.request;
import cn.hutool.core.lang.Assert;
import cn.hutool.core.net.url.UrlQuery;
@ -7,8 +7,8 @@ import cn.hutool.core.util.StrUtil;
import com.aizuda.snailjob.common.core.constant.SystemConstants;
import com.aizuda.snailjob.common.core.constant.SystemConstants.HTTP_PATH;
import com.aizuda.snailjob.common.core.enums.StatusEnum;
import com.aizuda.snailjob.common.core.model.NettyResult;
import com.aizuda.snailjob.common.core.model.SnailJobRequest;
import com.aizuda.snailjob.common.core.model.SnailJobRpcResult;
import com.aizuda.snailjob.common.core.util.JsonUtil;
import com.aizuda.snailjob.common.log.SnailJobLog;
import com.aizuda.snailjob.server.common.WaitStrategy;
@ -51,7 +51,7 @@ public class OpenApiAddJobRequestHandler extends PostHttpRequestHandler {
}
@Override
public String doHandler(String content, UrlQuery query, HttpHeaders headers) {
public SnailJobRpcResult doHandler(String content, UrlQuery query, HttpHeaders headers) {
SnailJobLog.LOCAL.debug("Add job content:[{}]", content);
SnailJobRequest retryRequest = JsonUtil.parseObject(content, SnailJobRequest.class);
Object[] args = retryRequest.getArgs();
@ -68,7 +68,7 @@ public class OpenApiAddJobRequestHandler extends PostHttpRequestHandler {
job.setNamespaceId(HttpHeaderUtil.getNamespace(headers));
job.setId(null);
Assert.isTrue(1 == jobMapper.insert(job), ()-> new SnailJobServerException("新增任务失败"));
return JsonUtil.toJsonString(new NettyResult(job.getId(), retryRequest.getReqId()));
return new SnailJobRpcResult(job.getId(), retryRequest.getReqId());
}
private Integer isResident(JobRequestVO jobRequestVO) {

View File

@ -1,10 +1,10 @@
package com.aizuda.snailjob.server.job.task.support.handler;
package com.aizuda.snailjob.server.job.task.support.request;
import cn.hutool.core.lang.Assert;
import cn.hutool.core.net.url.UrlQuery;
import com.aizuda.snailjob.common.core.constant.SystemConstants.HTTP_PATH;
import com.aizuda.snailjob.common.core.model.NettyResult;
import com.aizuda.snailjob.common.core.model.SnailJobRequest;
import com.aizuda.snailjob.common.core.model.SnailJobRpcResult;
import com.aizuda.snailjob.common.core.util.JsonUtil;
import com.aizuda.snailjob.common.log.SnailJobLog;
import com.aizuda.snailjob.server.common.convert.JobResponseVOConverter;
@ -38,7 +38,7 @@ public class OpenApiGetJobDetailRequestHandler extends PostHttpRequestHandler {
}
@Override
public String doHandler(String content, UrlQuery query, HttpHeaders headers) {
public SnailJobRpcResult doHandler(String content, UrlQuery query, HttpHeaders headers) {
SnailJobLog.LOCAL.debug("Update job content:[{}]", content);
SnailJobRequest retryRequest = JsonUtil.parseObject(content, SnailJobRequest.class);
Object[] args = retryRequest.getArgs();
@ -47,7 +47,7 @@ public class OpenApiGetJobDetailRequestHandler extends PostHttpRequestHandler {
Job job = jobMapper.selectById(jobId);
JobResponseVO convert = JobResponseVOConverter.INSTANCE.convert(job);
return JsonUtil.toJsonString(new NettyResult(convert, retryRequest.getReqId()));
return new SnailJobRpcResult(convert, retryRequest.getReqId());
}
}

View File

@ -1,10 +1,10 @@
package com.aizuda.snailjob.server.job.task.support.handler;
package com.aizuda.snailjob.server.job.task.support.request;
import cn.hutool.core.net.url.UrlQuery;
import com.aizuda.snailjob.common.core.constant.SystemConstants.HTTP_PATH;
import com.aizuda.snailjob.common.core.enums.StatusEnum;
import com.aizuda.snailjob.common.core.model.NettyResult;
import com.aizuda.snailjob.common.core.model.SnailJobRequest;
import com.aizuda.snailjob.common.core.model.SnailJobRpcResult;
import com.aizuda.snailjob.common.core.util.JsonUtil;
import com.aizuda.snailjob.common.log.SnailJobLog;
import com.aizuda.snailjob.server.common.enums.JobTaskExecutorSceneEnum;
@ -47,7 +47,7 @@ public class OpenApiTriggerJobRequestHandler extends PostHttpRequestHandler {
}
@Override
public String doHandler(String content, UrlQuery query, HttpHeaders headers) {
public SnailJobRpcResult doHandler(String content, UrlQuery query, HttpHeaders headers) {
SnailJobLog.LOCAL.debug("Trigger job content:[{}]", content);
SnailJobRequest retryRequest = JsonUtil.parseObject(content, SnailJobRequest.class);
Object[] args = retryRequest.getArgs();
@ -55,7 +55,7 @@ public class OpenApiTriggerJobRequestHandler extends PostHttpRequestHandler {
Job job = jobMapper.selectById(jobId);
if (Objects.isNull(job)){
SnailJobLog.LOCAL.warn("job can not be null.");
return JsonUtil.toJsonString(new NettyResult(false, retryRequest.getReqId()));
return new SnailJobRpcResult(false, retryRequest.getReqId());
}
long count = accessTemplate.getGroupConfigAccess().count(new LambdaQueryWrapper<GroupConfig>()
@ -66,7 +66,7 @@ public class OpenApiTriggerJobRequestHandler extends PostHttpRequestHandler {
if (count <= 0){
SnailJobLog.LOCAL.warn("组:[{}]已经关闭,不支持手动执行.", job.getGroupName());
return JsonUtil.toJsonString(new NettyResult(false, retryRequest.getReqId()));
return new SnailJobRpcResult(false, retryRequest.getReqId());
}
JobTaskPrepareDTO jobTaskPrepare = JobTaskConverter.INSTANCE.toJobTaskPrepare(job);
// 设置now表示立即执行
@ -75,6 +75,6 @@ public class OpenApiTriggerJobRequestHandler extends PostHttpRequestHandler {
// 创建批次
terminalJobPrepareHandler.handle(jobTaskPrepare);
return JsonUtil.toJsonString(new NettyResult(true, retryRequest.getReqId()));
return new SnailJobRpcResult(true, retryRequest.getReqId());
}
}

View File

@ -1,12 +1,12 @@
package com.aizuda.snailjob.server.job.task.support.handler;
package com.aizuda.snailjob.server.job.task.support.request;
import cn.hutool.core.collection.CollectionUtil;
import cn.hutool.core.net.url.UrlQuery;
import cn.hutool.core.util.StrUtil;
import com.aizuda.snailjob.common.core.constant.SystemConstants.HTTP_PATH;
import com.aizuda.snailjob.common.core.enums.StatusEnum;
import com.aizuda.snailjob.common.core.model.NettyResult;
import com.aizuda.snailjob.common.core.model.SnailJobRequest;
import com.aizuda.snailjob.common.core.model.SnailJobRpcResult;
import com.aizuda.snailjob.common.core.util.JsonUtil;
import com.aizuda.snailjob.common.log.SnailJobLog;
import com.aizuda.snailjob.server.common.enums.JobTaskExecutorSceneEnum;
@ -52,7 +52,7 @@ public class OpenApiTriggerWorkFlowRequestHandler extends PostHttpRequestHandler
}
@Override
public String doHandler(String content, UrlQuery query, HttpHeaders headers) {
public SnailJobRpcResult doHandler(String content, UrlQuery query, HttpHeaders headers) {
SnailJobLog.LOCAL.debug("Trigger job content:[{}]", content);
SnailJobRequest retryRequest = JsonUtil.parseObject(content, SnailJobRequest.class);
Object[] args = retryRequest.getArgs();
@ -60,7 +60,7 @@ public class OpenApiTriggerWorkFlowRequestHandler extends PostHttpRequestHandler
Workflow workflow = workflowMapper.selectById(id);
if (Objects.isNull(workflow)){
SnailJobLog.LOCAL.warn("workflow can not be null.");
return JsonUtil.toJsonString(new NettyResult(false, retryRequest.getReqId()));
return new SnailJobRpcResult(false, retryRequest.getReqId());
}
// 将字符串反序列化为 Set
if (StrUtil.isNotBlank(workflow.getGroupName())) {
@ -78,7 +78,7 @@ public class OpenApiTriggerWorkFlowRequestHandler extends PostHttpRequestHandler
if (count <= 0){
SnailJobLog.LOCAL.warn("组:[{}]已经关闭,不支持手动执行.", workflow.getGroupName());
return JsonUtil.toJsonString(new NettyResult(false, retryRequest.getReqId()));
return new SnailJobRpcResult(false, retryRequest.getReqId());
}
}
}
@ -91,7 +91,7 @@ public class OpenApiTriggerWorkFlowRequestHandler extends PostHttpRequestHandler
terminalWorkflowPrepareHandler.handler(prepareDTO);
return JsonUtil.toJsonString(new NettyResult(true, retryRequest.getReqId()));
return new SnailJobRpcResult(true, retryRequest.getReqId());
}
}

View File

@ -1,11 +1,11 @@
package com.aizuda.snailjob.server.job.task.support.handler;
package com.aizuda.snailjob.server.job.task.support.request;
import cn.hutool.core.net.url.UrlQuery;
import com.aizuda.snailjob.common.core.constant.SystemConstants;
import com.aizuda.snailjob.common.core.constant.SystemConstants.HTTP_PATH;
import com.aizuda.snailjob.common.core.enums.StatusEnum;
import com.aizuda.snailjob.common.core.model.NettyResult;
import com.aizuda.snailjob.common.core.model.SnailJobRequest;
import com.aizuda.snailjob.common.core.model.SnailJobRpcResult;
import com.aizuda.snailjob.common.core.util.JsonUtil;
import com.aizuda.snailjob.common.log.SnailJobLog;
import com.aizuda.snailjob.server.common.WaitStrategy;
@ -48,7 +48,7 @@ public class OpenApiUpdateJobRequestHandler extends PostHttpRequestHandler {
}
@Override
public String doHandler(String content, UrlQuery query, HttpHeaders headers) {
public SnailJobRpcResult doHandler(String content, UrlQuery query, HttpHeaders headers) {
SnailJobLog.LOCAL.debug("Update job content:[{}]", content);
SnailJobRequest retryRequest = JsonUtil.parseObject(content, SnailJobRequest.class);
Object[] args = retryRequest.getArgs();
@ -56,13 +56,13 @@ public class OpenApiUpdateJobRequestHandler extends PostHttpRequestHandler {
JobRequestVO jobRequestVO = JsonUtil.parseObject(JsonUtil.toJsonString(args[0]), JobRequestVO.class);
if (Objects.isNull(jobRequestVO.getId())){
SnailJobLog.LOCAL.warn("id不能为空更新失败");
return JsonUtil.toJsonString(new NettyResult(false, retryRequest.getReqId()));
return new SnailJobRpcResult(false, retryRequest.getReqId());
}
Job job = jobMapper.selectById(jobRequestVO.getId());
if (Objects.isNull(job)){
SnailJobLog.LOCAL.warn("job为空更新失败");
return JsonUtil.toJsonString(new NettyResult(false, retryRequest.getReqId()));
return new SnailJobRpcResult(false, retryRequest.getReqId());
}
// 判断常驻任务
@ -93,7 +93,7 @@ public class OpenApiUpdateJobRequestHandler extends PostHttpRequestHandler {
// 禁止更新组
updateJob.setGroupName(null);
boolean insert = 1 == jobMapper.updateById(updateJob);
return JsonUtil.toJsonString(new NettyResult(insert, retryRequest.getReqId()));
return new SnailJobRpcResult(insert, retryRequest.getReqId());
}

View File

@ -1,9 +1,9 @@
package com.aizuda.snailjob.server.job.task.support.handler;
package com.aizuda.snailjob.server.job.task.support.request;
import cn.hutool.core.net.url.UrlQuery;
import com.aizuda.snailjob.common.core.constant.SystemConstants.HTTP_PATH;
import com.aizuda.snailjob.common.core.model.NettyResult;
import com.aizuda.snailjob.common.core.model.SnailJobRequest;
import com.aizuda.snailjob.common.core.model.SnailJobRpcResult;
import com.aizuda.snailjob.common.core.util.JsonUtil;
import com.aizuda.snailjob.common.log.SnailJobLog;
import com.aizuda.snailjob.server.common.handler.PostHttpRequestHandler;
@ -36,20 +36,20 @@ public class OpenApiUpdateJobStatusRequestHandler extends PostHttpRequestHandler
}
@Override
public String doHandler(String content, UrlQuery query, HttpHeaders headers) {
public SnailJobRpcResult doHandler(String content, UrlQuery query, HttpHeaders headers) {
SnailJobRequest retryRequest = JsonUtil.parseObject(content, SnailJobRequest.class);
Object[] args = retryRequest.getArgs();
JobStatusUpdateRequestVO jobRequestVO = JsonUtil.parseObject(JsonUtil.toJsonString(args[0]), JobStatusUpdateRequestVO.class);
Long count = jobMapper.selectCount(new LambdaQueryWrapper<Job>().eq(Job::getId, jobRequestVO.getId()));
if (1 != count){
SnailJobLog.LOCAL.warn("更新任务失败");
return JsonUtil.toJsonString(new NettyResult(false, retryRequest.getReqId()));
return new SnailJobRpcResult(false, retryRequest.getReqId());
}
Job job = new Job();
job.setId(jobRequestVO.getId());
job.setJobStatus(jobRequestVO.getJobStatus());
boolean update = 1 == jobMapper.updateById(job);
return JsonUtil.toJsonString(new NettyResult(update, retryRequest.getReqId()));
return new SnailJobRpcResult(update, retryRequest.getReqId());
}
}

View File

@ -1,9 +1,9 @@
package com.aizuda.snailjob.server.job.task.support.handler;
package com.aizuda.snailjob.server.job.task.support.request;
import cn.hutool.core.net.url.UrlQuery;
import com.aizuda.snailjob.common.core.constant.SystemConstants.HTTP_PATH;
import com.aizuda.snailjob.common.core.model.NettyResult;
import com.aizuda.snailjob.common.core.model.SnailJobRequest;
import com.aizuda.snailjob.common.core.model.SnailJobRpcResult;
import com.aizuda.snailjob.common.core.util.JsonUtil;
import com.aizuda.snailjob.common.log.SnailJobLog;
import com.aizuda.snailjob.server.common.handler.PostHttpRequestHandler;
@ -38,7 +38,7 @@ public class OpenApiUpdateWorkFlowStatusRequestHandler extends PostHttpRequestHa
}
@Override
public String doHandler(String content, UrlQuery query, HttpHeaders headers) {
public SnailJobRpcResult doHandler(String content, UrlQuery query, HttpHeaders headers) {
SnailJobRequest retryRequest = JsonUtil.parseObject(content, SnailJobRequest.class);
Object[] args = retryRequest.getArgs();
JobStatusUpdateRequestVO jobRequestVO = JsonUtil.parseObject(JsonUtil.toJsonString(args[0]), JobStatusUpdateRequestVO.class);
@ -49,12 +49,12 @@ public class OpenApiUpdateWorkFlowStatusRequestHandler extends PostHttpRequestHa
if (Objects.isNull(workflow)){
SnailJobLog.LOCAL.warn("工作流不存在");
return JsonUtil.toJsonString(new NettyResult(false, retryRequest.getReqId()));
return new SnailJobRpcResult(false, retryRequest.getReqId());
}
workflow.setWorkflowStatus(jobRequestVO.getJobStatus());
boolean update = 1 == workflowMapper.updateById(workflow);
return JsonUtil.toJsonString(new NettyResult(update, retryRequest.getReqId()));
return new SnailJobRpcResult(update, retryRequest.getReqId());
}
}