feat(1.5.0-beta1): 新增sceneId和groupId

This commit is contained in:
opensnail 2025-04-09 23:24:14 +08:00
parent cf91d88ac2
commit 07fe8283a9
16 changed files with 47 additions and 23 deletions

View File

@ -3,14 +3,16 @@
<mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper"> <mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper">
<insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id"> <insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id">
INSERT INTO sj_retry_dead_letter (namespace_id, group_name, scene_name, idempotent_id, biz_no, INSERT INTO sj_retry_dead_letter (namespace_id, group_name, group_id, scene_name, scene_id, idempotent_id, biz_no,
executor_name, args_str, ext_attrs, create_dt) executor_name, args_str, ext_attrs, create_dt)
VALUES VALUES
<foreach collection="list" item="item" separator=","> <foreach collection="list" item="item" separator=",">
( (
#{item.namespaceId,jdbcType=VARCHAR}, #{item.namespaceId,jdbcType=VARCHAR},
#{item.groupName,jdbcType=VARCHAR}, #{item.groupName,jdbcType=VARCHAR},
#{item.groupId,jdbcType=BIGINT},
#{item.sceneName,jdbcType=VARCHAR}, #{item.sceneName,jdbcType=VARCHAR},
#{item.sceneId,jdbcType=BIGINT},
#{item.idempotentId,jdbcType=VARCHAR}, #{item.idempotentId,jdbcType=VARCHAR},
#{item.bizNo,jdbcType=VARCHAR}, #{item.bizNo,jdbcType=VARCHAR},
#{item.executorName,jdbcType=VARCHAR}, #{item.executorName,jdbcType=VARCHAR},

View File

@ -3,16 +3,16 @@
<mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryMapper"> <mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryMapper">
<insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id"> <insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id">
INSERT INTO sj_retry (namespace_id, group_name, scene_name, idempotent_id, biz_no, INSERT INTO sj_retry (namespace_id, group_name, group_id, scene_name, scene_id, idempotent_id, biz_no,
executor_name, args_str, ext_attrs, next_trigger_at, task_type, retry_status, create_dt, executor_name, args_str, ext_attrs, next_trigger_at, task_type, retry_status, create_dt,
bucket_index, parent_id, deleted) bucket_index, parent_id, deleted)
VALUES VALUES
<foreach collection="list" item="item" separator=","> <foreach collection="list" item="item" separator=",">
( (
#{item.namespaceId}, #{item.groupName}, #{item.sceneName}, #{item.idempotentId}, #{item.namespaceId}, #{item.groupName},#{item.groupId}, #{item.sceneName}, #{item.sceneId},
#{item.bizNo}, #{item.executorName}, #{item.argsStr}, #{item.extAttrs}, #{item.nextTriggerAt}, #{item.idempotentId}, #{item.bizNo}, #{item.executorName}, #{item.argsStr}, #{item.extAttrs},
#{item.taskType}, #{item.retryStatus}, #{item.createDt}, #{item.bucketIndex}, #{item.parentId}, #{item.nextTriggerAt}, #{item.taskType}, #{item.retryStatus}, #{item.createDt}, #{item.bucketIndex},
#{item.deleted} #{item.parentId}, #{item.deleted}
) )
</foreach> </foreach>
</insert> </insert>

View File

@ -3,14 +3,16 @@
<mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper"> <mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper">
<insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id"> <insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id">
INSERT INTO sj_retry_dead_letter (namespace_id, group_name, scene_name, idempotent_id, biz_no, INSERT INTO sj_retry_dead_letter (namespace_id, group_name, group_id, scene_name, scene_id, idempotent_id, biz_no,
executor_name, args_str, ext_attrs, create_dt) executor_name, args_str, ext_attrs, create_dt)
VALUES VALUES
<foreach collection="list" item="item" separator=","> <foreach collection="list" item="item" separator=",">
( (
#{item.namespaceId,jdbcType=VARCHAR}, #{item.namespaceId,jdbcType=VARCHAR},
#{item.groupName,jdbcType=VARCHAR}, #{item.groupName,jdbcType=VARCHAR},
#{item.groupId,jdbcType=BIGINT},
#{item.sceneName,jdbcType=VARCHAR}, #{item.sceneName,jdbcType=VARCHAR},
#{item.sceneId,jdbcType=BIGINT},
#{item.idempotentId,jdbcType=VARCHAR}, #{item.idempotentId,jdbcType=VARCHAR},
#{item.bizNo,jdbcType=VARCHAR}, #{item.bizNo,jdbcType=VARCHAR},
#{item.executorName,jdbcType=VARCHAR}, #{item.executorName,jdbcType=VARCHAR},

View File

@ -4,14 +4,14 @@
<!-- 定义批量新增的 SQL 映射 --> <!-- 定义批量新增的 SQL 映射 -->
<insert id="insertBatch" parameterType="java.util.List"> <insert id="insertBatch" parameterType="java.util.List">
INSERT INTO sj_retry (namespace_id, group_name, scene_name, INSERT INTO sj_retry (namespace_id, group_name, group_id, scene_name, scene_id,
idempotent_id, biz_no, executor_name, args_str, ext_attrs, idempotent_id, biz_no, executor_name, args_str, ext_attrs,
next_trigger_at, task_type, retry_status, create_dt, bucket_index, parent_id, deleted) next_trigger_at, task_type, retry_status, create_dt, bucket_index, parent_id, deleted)
VALUES VALUES
<foreach collection="list" item="item" separator=","> <foreach collection="list" item="item" separator=",">
( (
#{item.namespaceId}, #{item.groupName}, #{item.namespaceId}, #{item.groupName},#{item.groupId}, #{item.sceneName}, #{item.sceneId},
#{item.sceneName}, #{item.idempotentId}, #{item.bizNo}, #{item.executorName}, #{item.argsStr}, #{item.idempotentId}, #{item.bizNo}, #{item.executorName}, #{item.argsStr},
#{item.extAttrs}, #{item.nextTriggerAt}, #{item.taskType}, #{item.retryStatus}, #{item.createDt}, #{item.extAttrs}, #{item.nextTriggerAt}, #{item.taskType}, #{item.retryStatus}, #{item.createDt},
#{item.bucketIndex}, #{item.parentId}, #{item.bucketIndex}, #{item.parentId},
#{item.deleted} #{item.deleted}

View File

@ -3,14 +3,16 @@
<mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper"> <mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper">
<insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id"> <insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id">
INSERT INTO sj_retry_dead_letter (namespace_id, group_name, scene_name, idempotent_id, biz_no, INSERT INTO sj_retry_dead_letter (namespace_id, group_name, group_id, scene_name, scene_id, idempotent_id, biz_no,
executor_name, args_str, ext_attrs, create_dt) executor_name, args_str, ext_attrs, create_dt)
VALUES VALUES
<foreach collection="list" item="item" separator=","> <foreach collection="list" item="item" separator=",">
( (
#{item.namespaceId,jdbcType=VARCHAR}, #{item.namespaceId,jdbcType=VARCHAR},
#{item.groupName,jdbcType=VARCHAR}, #{item.groupName,jdbcType=VARCHAR},
#{item.groupId,jdbcType=BIGINT},
#{item.sceneName,jdbcType=VARCHAR}, #{item.sceneName,jdbcType=VARCHAR},
#{item.sceneId,jdbcType=BIGINT},
#{item.idempotentId,jdbcType=VARCHAR}, #{item.idempotentId,jdbcType=VARCHAR},
#{item.bizNo,jdbcType=VARCHAR}, #{item.bizNo,jdbcType=VARCHAR},
#{item.executorName,jdbcType=VARCHAR}, #{item.executorName,jdbcType=VARCHAR},

View File

@ -3,7 +3,7 @@
<mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryMapper"> <mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryMapper">
<insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id"> <insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id">
INSERT INTO sj_retry (namespace_id, group_name, scene_name, idempotent_id, biz_no, INSERT INTO sj_retry (namespace_id, group_name, group_id, scene_name, scene_id, idempotent_id, biz_no,
executor_name, args_str, ext_attrs, next_trigger_at, task_type, retry_status, executor_name, args_str, ext_attrs, next_trigger_at, task_type, retry_status,
create_dt, bucket_index, parent_id, deleted) create_dt, bucket_index, parent_id, deleted)
VALUES VALUES
@ -11,7 +11,9 @@
( (
#{item.namespaceId}, #{item.namespaceId},
#{item.groupName}, #{item.groupName},
#{item.groupId},
#{item.sceneName}, #{item.sceneName},
#{item.sceneId},
#{item.idempotentId}, #{item.idempotentId},
#{item.bizNo}, #{item.bizNo},
#{item.executorName}, #{item.executorName},

View File

@ -3,14 +3,16 @@
<mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper"> <mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper">
<insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id"> <insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id">
INSERT INTO sj_retry_dead_letter (namespace_id, group_name, scene_name, idempotent_id, biz_no, INSERT INTO sj_retry_dead_letter (namespace_id, group_name, group_id, scene_name, scene_id, idempotent_id, biz_no,
executor_name, args_str, ext_attrs, create_dt) executor_name, args_str, ext_attrs, create_dt)
VALUES VALUES
<foreach collection="list" item="item" separator=","> <foreach collection="list" item="item" separator=",">
( (
#{item.namespaceId,jdbcType=VARCHAR}, #{item.namespaceId,jdbcType=VARCHAR},
#{item.groupName,jdbcType=VARCHAR}, #{item.groupName,jdbcType=VARCHAR},
#{item.groupId,jdbcType=BIGINT},
#{item.sceneName,jdbcType=VARCHAR}, #{item.sceneName,jdbcType=VARCHAR},
#{item.sceneId,jdbcType=BIGINT},
#{item.idempotentId,jdbcType=VARCHAR}, #{item.idempotentId,jdbcType=VARCHAR},
#{item.bizNo,jdbcType=VARCHAR}, #{item.bizNo,jdbcType=VARCHAR},
#{item.executorName,jdbcType=VARCHAR}, #{item.executorName,jdbcType=VARCHAR},

View File

@ -3,7 +3,7 @@
<mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryMapper"> <mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryMapper">
<insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id"> <insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id">
INSERT INTO sj_retry (namespace_id, group_name, scene_name, idempotent_id, biz_no, INSERT INTO sj_retry (namespace_id, group_name, group_id, scene_name, scene_id, idempotent_id, biz_no,
executor_name, args_str, ext_attrs, next_trigger_at, task_type, retry_status, executor_name, args_str, ext_attrs, next_trigger_at, task_type, retry_status,
create_dt, bucket_index, parent_id, deleted) create_dt, bucket_index, parent_id, deleted)
VALUES VALUES
@ -11,7 +11,9 @@
( (
#{item.namespaceId}, #{item.namespaceId},
#{item.groupName}, #{item.groupName},
#{item.groupId},
#{item.sceneName}, #{item.sceneName},
#{item.sceneId},
#{item.idempotentId}, #{item.idempotentId},
#{item.bizNo}, #{item.bizNo},
#{item.executorName}, #{item.executorName},

View File

@ -3,13 +3,15 @@
<mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper"> <mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper">
<insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="false"> <insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="false">
INSERT INTO sj_retry_dead_letter (namespace_id, group_name, scene_name, idempotent_id, biz_no, INSERT INTO sj_retry_dead_letter (namespace_id, group_name, group_id, scene_name, scene_id, idempotent_id, biz_no,
executor_name, args_str, ext_attrs, create_dt) executor_name, args_str, ext_attrs, create_dt)
<foreach collection="list" item="item" separator="UNION ALL"> <foreach collection="list" item="item" separator="UNION ALL">
SELECT SELECT
#{item.namespaceId,jdbcType=VARCHAR}, #{item.namespaceId,jdbcType=VARCHAR},
#{item.groupName,jdbcType=VARCHAR}, #{item.groupName,jdbcType=VARCHAR},
#{item.groupId,jdbcType=BIGINT},
#{item.sceneName,jdbcType=VARCHAR}, #{item.sceneName,jdbcType=VARCHAR},
#{item.sceneId,jdbcType=BIGINT},
#{item.idempotentId,jdbcType=VARCHAR}, #{item.idempotentId,jdbcType=VARCHAR},
#{item.bizNo,jdbcType=VARCHAR}, #{item.bizNo,jdbcType=VARCHAR},
#{item.executorName,jdbcType=VARCHAR}, #{item.executorName,jdbcType=VARCHAR},

View File

@ -3,12 +3,13 @@
<mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryMapper"> <mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryMapper">
<insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="false"> <insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="false">
INSERT INTO sj_retry (namespace_id, group_name, scene_name, idempotent_id, biz_no, INSERT INTO sj_retry (namespace_id, group_name, group_id, scene_name, scene_id, idempotent_id, biz_no,
executor_name, args_str, ext_attrs, next_trigger_at, task_type, retry_status, create_dt executor_name, args_str, ext_attrs, next_trigger_at, task_type, retry_status, create_dt
, bucket_index, parent_id, deleted) , bucket_index, parent_id, deleted)
<foreach collection="list" item="item" separator="UNION ALL"> <foreach collection="list" item="item" separator="UNION ALL">
SELECT SELECT
#{item.namespaceId}, #{item.groupName}, #{item.sceneName}, #{item.idempotentId}, #{item.namespaceId}, #{item.groupName}, #{item.groupId}, #{item.sceneName}, #{item.sceneId},
#{item.idempotentId},
#{item.bizNo}, #{item.executorName}, #{item.argsStr}, #{item.extAttrs}, #{item.nextTriggerAt}, #{item.bizNo}, #{item.executorName}, #{item.argsStr}, #{item.extAttrs}, #{item.nextTriggerAt},
#{item.taskType}, #{item.retryStatus}, #{item.createDt}, #{item.bucketIndex}, #{item.parentId}, #{item.taskType}, #{item.retryStatus}, #{item.createDt}, #{item.bucketIndex}, #{item.parentId},
#{item.deleted} #{item.deleted}

View File

@ -3,14 +3,16 @@
<mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper"> <mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper">
<insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id"> <insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id">
INSERT INTO sj_retry_dead_letter (namespace_id, group_name, scene_name, idempotent_id, biz_no, INSERT INTO sj_retry_dead_letter (namespace_id, group_name, group_id, scene_name, scene_id, idempotent_id, biz_no,
executor_name, args_str, ext_attrs, create_dt) executor_name, args_str, ext_attrs, create_dt)
VALUES VALUES
<foreach collection="list" item="item" separator=","> <foreach collection="list" item="item" separator=",">
( (
#{item.namespaceId,jdbcType=VARCHAR}, #{item.namespaceId,jdbcType=VARCHAR},
#{item.groupName,jdbcType=VARCHAR}, #{item.groupName,jdbcType=VARCHAR},
#{item.groupId,jdbcType=BIGINT},
#{item.sceneName,jdbcType=VARCHAR}, #{item.sceneName,jdbcType=VARCHAR},
#{item.sceneId,jdbcType=BIGINT},
#{item.idempotentId,jdbcType=VARCHAR}, #{item.idempotentId,jdbcType=VARCHAR},
#{item.bizNo,jdbcType=VARCHAR}, #{item.bizNo,jdbcType=VARCHAR},
#{item.executorName,jdbcType=VARCHAR}, #{item.executorName,jdbcType=VARCHAR},

View File

@ -4,14 +4,14 @@
<!-- 定义批量新增的 SQL 映射 --> <!-- 定义批量新增的 SQL 映射 -->
<insert id="insertBatch" parameterType="java.util.List"> <insert id="insertBatch" parameterType="java.util.List">
INSERT INTO sj_retry (namespace_id, group_name, scene_name, INSERT INTO sj_retry (namespace_id, group_name, group_id, scene_name, scene_id,
idempotent_id, biz_no, executor_name, args_str, ext_attrs, idempotent_id, biz_no, executor_name, args_str, ext_attrs,
next_trigger_at, task_type, retry_status, create_dt, bucket_index, parent_id, deleted) next_trigger_at, task_type, retry_status, create_dt, bucket_index, parent_id, deleted)
VALUES VALUES
<foreach collection="list" item="item" separator=","> <foreach collection="list" item="item" separator=",">
( (
#{item.namespaceId}, #{item.groupName},#{item.sceneName}, #{item.idempotentId}, #{item.bizNo}, #{item.namespaceId}, #{item.groupName}, #{item.groupId}, #{item.sceneName}, #{item.sceneId}, #{item.idempotentId},
#{item.executorName}, #{item.argsStr}, #{item.extAttrs}, #{item.nextTriggerAt}, #{item.taskType}, #{item.bizNo},#{item.executorName}, #{item.argsStr}, #{item.extAttrs}, #{item.nextTriggerAt}, #{item.taskType},
#{item.retryStatus}, #{item.createDt}, #{item.bucketIndex}, #{item.parentId}, #{item.deleted} #{item.retryStatus}, #{item.createDt}, #{item.bucketIndex}, #{item.parentId}, #{item.deleted}
) )
</foreach> </foreach>

View File

@ -3,14 +3,16 @@
<mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper"> <mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryDeadLetterMapper">
<insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id"> <insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id">
INSERT INTO sj_retry_dead_letter (namespace_id, group_name, scene_name, idempotent_id, biz_no, INSERT INTO sj_retry_dead_letter (namespace_id, group_name, group_id, scene_name, scene_id, idempotent_id, biz_no,
executor_name, args_str, ext_attrs, create_dt) executor_name, args_str, ext_attrs, create_dt)
VALUES VALUES
<foreach collection="list" item="item" separator=","> <foreach collection="list" item="item" separator=",">
( (
#{item.namespaceId,jdbcType=VARCHAR}, #{item.namespaceId,jdbcType=VARCHAR},
#{item.groupName,jdbcType=VARCHAR}, #{item.groupName,jdbcType=VARCHAR},
#{item.groupId,jdbcType=BIGINT},
#{item.sceneName,jdbcType=VARCHAR}, #{item.sceneName,jdbcType=VARCHAR},
#{item.sceneId,jdbcType=BIGINT},
#{item.idempotentId,jdbcType=VARCHAR}, #{item.idempotentId,jdbcType=VARCHAR},
#{item.bizNo,jdbcType=VARCHAR}, #{item.bizNo,jdbcType=VARCHAR},
#{item.executorName,jdbcType=VARCHAR}, #{item.executorName,jdbcType=VARCHAR},

View File

@ -3,7 +3,7 @@
<mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryMapper"> <mapper namespace="com.aizuda.snailjob.template.datasource.persistence.mapper.RetryMapper">
<insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id"> <insert id="insertBatch" parameterType="java.util.List" useGeneratedKeys="true" keyProperty="id">
INSERT INTO sj_retry (namespace_id, group_name, scene_name, idempotent_id, biz_no, INSERT INTO sj_retry (namespace_id, group_name, group_id, scene_name, scene_id, idempotent_id, biz_no,
executor_name, args_str, ext_attrs, next_trigger_at, task_type, retry_status, executor_name, args_str, ext_attrs, next_trigger_at, task_type, retry_status,
create_dt, bucket_index, parent_id, deleted) create_dt, bucket_index, parent_id, deleted)
VALUES VALUES
@ -11,7 +11,9 @@
( (
#{item.namespaceId}, #{item.namespaceId},
#{item.groupName}, #{item.groupName},
#{item.groupId},
#{item.sceneName}, #{item.sceneName},
#{item.sceneId},
#{item.idempotentId}, #{item.idempotentId},
#{item.bizNo}, #{item.bizNo},
#{item.executorName}, #{item.executorName},

View File

@ -125,6 +125,8 @@ public abstract class AbstractGenerator implements TaskGenerator {
retry.setGroupName(taskContext.getGroupName()); retry.setGroupName(taskContext.getGroupName());
retry.setSceneName(taskContext.getSceneName()); retry.setSceneName(taskContext.getSceneName());
retry.setRetryStatus(initStatus(taskContext)); retry.setRetryStatus(initStatus(taskContext));
retry.setSceneId(taskContext.getSceneId());
retry.setGroupId(taskContext.getGroupId());
retry.setParentId(0L); retry.setParentId(0L);
retry.setDeleted(0L); retry.setDeleted(0L);
if (StrUtil.isBlank(retry.getBizNo())) { if (StrUtil.isBlank(retry.getBizNo())) {

View File

@ -123,6 +123,7 @@ public class ReportRetryInfoHttpRequestHandler extends PostHttpRequestHandler {
TaskContext taskContext = new TaskContext(); TaskContext taskContext = new TaskContext();
taskContext.setSceneName(sceneName); taskContext.setSceneName(sceneName);
taskContext.setNamespaceId(namespaceId); taskContext.setNamespaceId(namespaceId);
taskContext.setGroupId(groupConfig.getId());
taskContext.setGroupName(groupName); taskContext.setGroupName(groupName);
taskContext.setInitScene(groupConfig.getInitScene()); taskContext.setInitScene(groupConfig.getInitScene());
taskContext.setTaskInfos(TaskContextConverter.INSTANCE.toTaskContextInfo(retryTaskDTOS)); taskContext.setTaskInfos(TaskContextConverter.INSTANCE.toTaskContextInfo(retryTaskDTOS));