Improvements to tests due to new architecture
This commit is contained in:
		
							parent
							
								
									d965df8c27
								
							
						
					
					
						commit
						53eb09bcb9
					
				@ -249,7 +249,7 @@ public class RuleChainActorMessageProcessor extends ComponentMsgProcessor<RuleCh
 | 
			
		||||
            if (relationsCount == 0) {
 | 
			
		||||
                log.trace("[{}][{}][{}] No outbound relations to process", tenantId, entityId, msg.getId());
 | 
			
		||||
                if (relationTypes.contains(TbRelationTypes.FAILURE)) {
 | 
			
		||||
                    RuleNodeCtx ruleNodeCtx =  nodeActors.get(originatorNodeId);
 | 
			
		||||
                    RuleNodeCtx ruleNodeCtx = nodeActors.get(originatorNodeId);
 | 
			
		||||
                    if (ruleNodeCtx != null) {
 | 
			
		||||
                        msg.getCallback().onFailure(new RuleNodeException(failureMessage, ruleChainName, ruleNodeCtx.getSelf()));
 | 
			
		||||
                    } else {
 | 
			
		||||
@ -327,6 +327,9 @@ public class RuleChainActorMessageProcessor extends ComponentMsgProcessor<RuleCh
 | 
			
		||||
    private void pushMsgToNode(RuleNodeCtx nodeCtx, TbMsg msg, String fromRelationType) {
 | 
			
		||||
        if (nodeCtx != null) {
 | 
			
		||||
            nodeCtx.getSelfActor().tell(new RuleChainToRuleNodeMsg(new DefaultTbContext(systemContext, nodeCtx), msg, fromRelationType), self);
 | 
			
		||||
        } else {
 | 
			
		||||
            log.error("[{}][{}] RuleNodeCtx is empty", entityId, ruleChainName);
 | 
			
		||||
            msg.getCallback().onFailure(new RuleEngineException("Rule Node CTX is empty"));
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@ -35,6 +35,7 @@ import org.thingsboard.server.common.data.id.DeviceId;
 | 
			
		||||
import org.thingsboard.server.common.data.id.RuleChainId;
 | 
			
		||||
import org.thingsboard.server.common.data.id.TenantId;
 | 
			
		||||
import org.thingsboard.server.common.data.rule.RuleChain;
 | 
			
		||||
import org.thingsboard.server.common.msg.MsgType;
 | 
			
		||||
import org.thingsboard.server.common.msg.TbActorMsg;
 | 
			
		||||
import org.thingsboard.server.common.msg.TbMsg;
 | 
			
		||||
import org.thingsboard.server.common.msg.aware.DeviceAwareMsg;
 | 
			
		||||
@ -73,24 +74,30 @@ public class TenantActor extends RuleChainManagerActor {
 | 
			
		||||
        log.info("[{}] Starting tenant actor.", tenantId);
 | 
			
		||||
        try {
 | 
			
		||||
            Tenant tenant = systemContext.getTenantService().findTenantById(tenantId);
 | 
			
		||||
            // This Service may be started for specific tenant only.
 | 
			
		||||
            Optional<TenantId> isolatedTenantId = systemContext.getServiceInfoProvider().getIsolatedTenant();
 | 
			
		||||
            if (tenant == null) {
 | 
			
		||||
                cantFindTenant = true;
 | 
			
		||||
                log.info("[{}] Started tenant actor for missing tenant.", tenantId);
 | 
			
		||||
            } else {
 | 
			
		||||
                // This Service may be started for specific tenant only.
 | 
			
		||||
                Optional<TenantId> isolatedTenantId = systemContext.getServiceInfoProvider().getIsolatedTenant();
 | 
			
		||||
 | 
			
		||||
            isRuleEngineForCurrentTenant = systemContext.getServiceInfoProvider().isService(ServiceType.TB_RULE_ENGINE);
 | 
			
		||||
            isCore = systemContext.getServiceInfoProvider().isService(ServiceType.TB_CORE);
 | 
			
		||||
                isRuleEngineForCurrentTenant = systemContext.getServiceInfoProvider().isService(ServiceType.TB_RULE_ENGINE);
 | 
			
		||||
                isCore = systemContext.getServiceInfoProvider().isService(ServiceType.TB_CORE);
 | 
			
		||||
 | 
			
		||||
            if (isRuleEngineForCurrentTenant) {
 | 
			
		||||
                try {
 | 
			
		||||
                    if (isolatedTenantId.map(id -> id.equals(tenantId)).orElseGet(() -> !tenant.isIsolatedTbRuleEngine())) {
 | 
			
		||||
                        initRuleChains();
 | 
			
		||||
                    } else {
 | 
			
		||||
                        isRuleEngineForCurrentTenant = false;
 | 
			
		||||
                if (isRuleEngineForCurrentTenant) {
 | 
			
		||||
                    try {
 | 
			
		||||
                        if (isolatedTenantId.map(id -> id.equals(tenantId)).orElseGet(() -> !tenant.isIsolatedTbRuleEngine())) {
 | 
			
		||||
                            log.info("[{}] Going to init rule chains", tenantId);
 | 
			
		||||
                            initRuleChains();
 | 
			
		||||
                        } else {
 | 
			
		||||
                            isRuleEngineForCurrentTenant = false;
 | 
			
		||||
                        }
 | 
			
		||||
                    } catch (Exception e) {
 | 
			
		||||
                        cantFindTenant = true;
 | 
			
		||||
                    }
 | 
			
		||||
                } catch (Exception e) {
 | 
			
		||||
                    cantFindTenant = true;
 | 
			
		||||
                }
 | 
			
		||||
                log.info("[{}] Tenant actor started.", tenantId);
 | 
			
		||||
            }
 | 
			
		||||
            log.info("[{}] Tenant actor started.", tenantId);
 | 
			
		||||
        } catch (Exception e) {
 | 
			
		||||
            log.warn("[{}] Unknown failure", tenantId, e);
 | 
			
		||||
        }
 | 
			
		||||
@ -104,7 +111,12 @@ public class TenantActor extends RuleChainManagerActor {
 | 
			
		||||
    @Override
 | 
			
		||||
    protected boolean process(TbActorMsg msg) {
 | 
			
		||||
        if (cantFindTenant) {
 | 
			
		||||
            log.info("Missing Tenant msg: {}", msg);
 | 
			
		||||
            log.info("[{}] Processing missing Tenant msg: {}", tenantId, msg);
 | 
			
		||||
            if (msg.getMsgType().equals(MsgType.QUEUE_TO_RULE_ENGINE_MSG)) {
 | 
			
		||||
                QueueToRuleEngineMsg queueMsg = (QueueToRuleEngineMsg) msg;
 | 
			
		||||
                queueMsg.getTbMsg().getCallback().onSuccess();
 | 
			
		||||
            }
 | 
			
		||||
            return true;
 | 
			
		||||
        }
 | 
			
		||||
        switch (msg.getMsgType()) {
 | 
			
		||||
            case PARTITION_CHANGE_MSG:
 | 
			
		||||
 | 
			
		||||
@ -344,7 +344,7 @@ public class TelemetryController extends BaseController {
 | 
			
		||||
        return deleteAttributes(entityId, scope, keysStr);
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    private DeferredResult<ResponseEntity> deleteAttributes(EntityId entityIdStr, String scope, String keysStr) throws ThingsboardException {
 | 
			
		||||
    private DeferredResult<ResponseEntity> deleteAttributes(EntityId entityIdSrc, String scope, String keysStr) throws ThingsboardException {
 | 
			
		||||
        List<String> keys = toKeysList(keysStr);
 | 
			
		||||
        if (keys.isEmpty()) {
 | 
			
		||||
            return getImmediateDeferredResult("Empty keys: " + keysStr, HttpStatus.BAD_REQUEST);
 | 
			
		||||
@ -354,13 +354,13 @@ public class TelemetryController extends BaseController {
 | 
			
		||||
        if (DataConstants.SERVER_SCOPE.equals(scope) ||
 | 
			
		||||
                DataConstants.SHARED_SCOPE.equals(scope) ||
 | 
			
		||||
                DataConstants.CLIENT_SCOPE.equals(scope)) {
 | 
			
		||||
            return accessValidator.validateEntityAndCallback(getCurrentUser(), Operation.WRITE_ATTRIBUTES, entityIdStr, (result, tenantId, entityId) -> {
 | 
			
		||||
            return accessValidator.validateEntityAndCallback(getCurrentUser(), Operation.WRITE_ATTRIBUTES, entityIdSrc, (result, tenantId, entityId) -> {
 | 
			
		||||
                ListenableFuture<List<Void>> future = attributesService.removeAll(user.getTenantId(), entityId, scope, keys);
 | 
			
		||||
                Futures.addCallback(future, new FutureCallback<List<Void>>() {
 | 
			
		||||
                    @Override
 | 
			
		||||
                    public void onSuccess(@Nullable List<Void> tmp) {
 | 
			
		||||
                        logAttributesDeleted(user, entityId, scope, keys, null);
 | 
			
		||||
                        if (entityId.getEntityType() == EntityType.DEVICE) {
 | 
			
		||||
                        if (entityIdSrc.getEntityType().equals(EntityType.DEVICE)) {
 | 
			
		||||
                            DeviceId deviceId = new DeviceId(entityId.getId());
 | 
			
		||||
                            Set<AttributeKey> keysToNotify = new HashSet<>();
 | 
			
		||||
                            keys.forEach(key -> keysToNotify.add(new AttributeKey(scope, key)));
 | 
			
		||||
@ -397,7 +397,7 @@ public class TelemetryController extends BaseController {
 | 
			
		||||
                    @Override
 | 
			
		||||
                    public void onSuccess(@Nullable Void tmp) {
 | 
			
		||||
                        logAttributesUpdated(user, entityId, scope, attributes, null);
 | 
			
		||||
                        if (entityId.getEntityType() == EntityType.DEVICE) {
 | 
			
		||||
                        if (entityIdSrc.getEntityType().equals(EntityType.DEVICE)) {
 | 
			
		||||
                            DeviceId deviceId = new DeviceId(entityId.getId());
 | 
			
		||||
                            tbClusterService.pushMsgToCore(DeviceAttributesEventNotificationMsg.onUpdate(
 | 
			
		||||
                                    user.getTenantId(), deviceId, scope, attributes), null);
 | 
			
		||||
 | 
			
		||||
@ -87,6 +87,7 @@ public class DefaultTbClusterService implements TbClusterService {
 | 
			
		||||
    @Override
 | 
			
		||||
    public void pushMsgToCore(ToDeviceActorNotificationMsg msg, TbQueueCallback callback) {
 | 
			
		||||
        TopicPartitionInfo tpi = partitionService.resolve(ServiceType.TB_CORE, msg.getTenantId(), msg.getDeviceId());
 | 
			
		||||
        log.trace("PUSHING msg: {} to:{}", msg, tpi);
 | 
			
		||||
        byte[] msgBytes = encodingService.encode(msg);
 | 
			
		||||
        ToCoreMsg toCoreMsg = ToCoreMsg.newBuilder().setToDeviceActorNotificationMsg(ByteString.copyFrom(msgBytes)).build();
 | 
			
		||||
        producerProvider.getTbCoreMsgProducer().send(tpi, new TbProtoQueueMsg<>(msg.getDeviceId().getId(), toCoreMsg), callback);
 | 
			
		||||
@ -96,6 +97,7 @@ public class DefaultTbClusterService implements TbClusterService {
 | 
			
		||||
    @Override
 | 
			
		||||
    public void pushNotificationToCore(String serviceId, FromDeviceRpcResponse response, TbQueueCallback callback) {
 | 
			
		||||
        TopicPartitionInfo tpi = partitionService.getNotificationsTopic(ServiceType.TB_CORE, serviceId);
 | 
			
		||||
        log.trace("PUSHING msg: {} to:{}", response, tpi);
 | 
			
		||||
        FromDeviceRPCResponseProto.Builder builder = FromDeviceRPCResponseProto.newBuilder()
 | 
			
		||||
                .setRequestIdMSB(response.getId().getMostSignificantBits())
 | 
			
		||||
                .setRequestIdLSB(response.getId().getLeastSignificantBits())
 | 
			
		||||
@ -108,6 +110,7 @@ public class DefaultTbClusterService implements TbClusterService {
 | 
			
		||||
 | 
			
		||||
    @Override
 | 
			
		||||
    public void pushMsgToRuleEngine(TopicPartitionInfo tpi, UUID msgId, ToRuleEngineMsg msg, TbQueueCallback callback) {
 | 
			
		||||
        log.trace("PUSHING msg: {} to:{}", msg, tpi);
 | 
			
		||||
        producerProvider.getRuleEngineMsgProducer().send(tpi, new TbProtoQueueMsg<>(msgId, msg), callback);
 | 
			
		||||
        toRuleEngineMsgs.incrementAndGet();
 | 
			
		||||
    }
 | 
			
		||||
@ -123,6 +126,7 @@ public class DefaultTbClusterService implements TbClusterService {
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
        TopicPartitionInfo tpi = partitionService.resolve(ServiceType.TB_RULE_ENGINE, tenantId, entityId);
 | 
			
		||||
        log.trace("PUSHING msg: {} to:{}", tbMsg, tpi);
 | 
			
		||||
        ToRuleEngineMsg msg = ToRuleEngineMsg.newBuilder()
 | 
			
		||||
                .setTenantIdMSB(tenantId.getId().getMostSignificantBits())
 | 
			
		||||
                .setTenantIdLSB(tenantId.getId().getLeastSignificantBits())
 | 
			
		||||
@ -134,6 +138,7 @@ public class DefaultTbClusterService implements TbClusterService {
 | 
			
		||||
    @Override
 | 
			
		||||
    public void pushNotificationToRuleEngine(String serviceId, FromDeviceRpcResponse response, TbQueueCallback callback) {
 | 
			
		||||
        TopicPartitionInfo tpi = partitionService.getNotificationsTopic(ServiceType.TB_RULE_ENGINE, serviceId);
 | 
			
		||||
        log.trace("PUSHING msg: {} to:{}", response, tpi);
 | 
			
		||||
        FromDeviceRPCResponseProto.Builder builder = FromDeviceRPCResponseProto.newBuilder()
 | 
			
		||||
                .setRequestIdMSB(response.getId().getMostSignificantBits())
 | 
			
		||||
                .setRequestIdLSB(response.getId().getLeastSignificantBits())
 | 
			
		||||
@ -147,6 +152,7 @@ public class DefaultTbClusterService implements TbClusterService {
 | 
			
		||||
    @Override
 | 
			
		||||
    public void pushNotificationToTransport(String serviceId, ToTransportMsg response, TbQueueCallback callback) {
 | 
			
		||||
        TopicPartitionInfo tpi = partitionService.getNotificationsTopic(ServiceType.TB_TRANSPORT, serviceId);
 | 
			
		||||
        log.trace("PUSHING msg: {} to:{}", response, tpi);
 | 
			
		||||
        producerProvider.getTransportNotificationsMsgProducer().send(tpi, new TbProtoQueueMsg<>(UUID.randomUUID(), response), callback);
 | 
			
		||||
        toTransportNfs.incrementAndGet();
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@ -174,7 +174,7 @@ public class DefaultTbRuleEngineConsumerService extends AbstractConsumerService<
 | 
			
		||||
                        }));
 | 
			
		||||
 | 
			
		||||
                        boolean timeout = false;
 | 
			
		||||
                        if (!ctx.await(packProcessingTimeout, TimeUnit.MILLISECONDS)) {
 | 
			
		||||
                        if (!ctx.await(configuration.getPackProcessingTimeout(), TimeUnit.MILLISECONDS)) {
 | 
			
		||||
                            timeout = true;
 | 
			
		||||
                        }
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@ -17,6 +17,8 @@ package org.thingsboard.server.service.queue.processing;
 | 
			
		||||
 | 
			
		||||
import lombok.extern.slf4j.Slf4j;
 | 
			
		||||
import org.springframework.stereotype.Component;
 | 
			
		||||
import org.thingsboard.server.common.msg.TbMsg;
 | 
			
		||||
import org.thingsboard.server.common.msg.queue.TbMsgCallback;
 | 
			
		||||
import org.thingsboard.server.gen.transport.TransportProtos;
 | 
			
		||||
import org.thingsboard.server.queue.common.TbProtoQueueMsg;
 | 
			
		||||
import org.thingsboard.server.queue.settings.TbRuleEngineQueueAckStrategyConfiguration;
 | 
			
		||||
@ -32,7 +34,7 @@ public class TbRuleEngineProcessingStrategyFactory {
 | 
			
		||||
 | 
			
		||||
    public TbRuleEngineProcessingStrategy newInstance(String name, TbRuleEngineQueueAckStrategyConfiguration configuration) {
 | 
			
		||||
        switch (configuration.getType()) {
 | 
			
		||||
            case "SKIP_ALL":
 | 
			
		||||
            case "SKIP_ALL_FAILURES":
 | 
			
		||||
                return new SkipStrategy(name);
 | 
			
		||||
            case "RETRY_ALL":
 | 
			
		||||
                return new RetryStrategy(name, true, true, true, configuration);
 | 
			
		||||
@ -98,7 +100,7 @@ public class TbRuleEngineProcessingStrategyFactory {
 | 
			
		||||
                    }
 | 
			
		||||
                    log.info("[{}] Going to reprocess {} messages", queueName, toReprocess.size());
 | 
			
		||||
                    if (log.isTraceEnabled()) {
 | 
			
		||||
                        toReprocess.forEach((id, msg) -> log.trace("Going to reprocess [{}]: {}", id, msg.getValue()));
 | 
			
		||||
                        toReprocess.forEach((id, msg) -> log.trace("Going to reprocess [{}]: {}", id, TbMsg.fromBytes(msg.getValue().getTbMsg().toByteArray(), TbMsgCallback.EMPTY)));
 | 
			
		||||
                    }
 | 
			
		||||
                    if (pauseBetweenRetries > 0) {
 | 
			
		||||
                        try {
 | 
			
		||||
@ -123,7 +125,15 @@ public class TbRuleEngineProcessingStrategyFactory {
 | 
			
		||||
 | 
			
		||||
        @Override
 | 
			
		||||
        public TbRuleEngineProcessingDecision analyze(TbRuleEngineProcessingResult result) {
 | 
			
		||||
            log.info("[{}] Reprocessing skipped for {} failed and {} timeout messages", queueName, result.getFailedMap().size(), result.getPendingMap().size());
 | 
			
		||||
            if (!result.isSuccess()) {
 | 
			
		||||
                log.info("[{}] Reprocessing skipped for {} failed and {} timeout messages", queueName, result.getFailedMap().size(), result.getPendingMap().size());
 | 
			
		||||
            }
 | 
			
		||||
            if (log.isTraceEnabled()) {
 | 
			
		||||
                result.getFailedMap().forEach((id, msg) -> log.trace("Failed messages [{}]: {}", id, TbMsg.fromBytes(msg.getValue().getTbMsg().toByteArray(), TbMsgCallback.EMPTY)));
 | 
			
		||||
            }
 | 
			
		||||
            if (log.isTraceEnabled()) {
 | 
			
		||||
                result.getPendingMap().forEach((id, msg) -> log.trace("Timeout messages [{}]: {}", id, TbMsg.fromBytes(msg.getValue().getTbMsg().toByteArray(), TbMsgCallback.EMPTY)));
 | 
			
		||||
            }
 | 
			
		||||
            return new TbRuleEngineProcessingDecision(true, null);
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@ -106,7 +106,7 @@ public class DefaultTbCoreDeviceRpcService implements TbCoreDeviceRpcService {
 | 
			
		||||
 | 
			
		||||
    @Override
 | 
			
		||||
    public void processRpcResponseFromRuleEngine(FromDeviceRpcResponse response) {
 | 
			
		||||
        log.trace("[{}] Received response to server-side RPC request from rule engine: [{}]", response.getId());
 | 
			
		||||
        log.trace("[{}] Received response to server-side RPC request from rule engine: [{}]", response.getId(), response);
 | 
			
		||||
        UUID requestId = response.getId();
 | 
			
		||||
        Consumer<FromDeviceRpcResponse> consumer = localToRuleEngineRpcRequests.remove(requestId);
 | 
			
		||||
        if (consumer != null) {
 | 
			
		||||
@ -177,9 +177,9 @@ public class DefaultTbCoreDeviceRpcService implements TbCoreDeviceRpcService {
 | 
			
		||||
 | 
			
		||||
    private void scheduleToRuleEngineTimeout(ToDeviceRpcRequest request, UUID requestId) {
 | 
			
		||||
        long timeout = Math.max(0, request.getExpirationTime() - System.currentTimeMillis());
 | 
			
		||||
        log.trace("[{}] processing to rule engine request: [{}]", this.hashCode(), requestId);
 | 
			
		||||
        log.trace("[{}] processing to rule engine request.", requestId);
 | 
			
		||||
        scheduler.schedule(() -> {
 | 
			
		||||
            log.trace("[{}] timeout for to rule engine request: [{}]", this.hashCode(), requestId);
 | 
			
		||||
            log.trace("[{}] timeout for processing to rule engine request.", requestId);
 | 
			
		||||
            Consumer<FromDeviceRpcResponse> consumer = localToRuleEngineRpcRequests.remove(requestId);
 | 
			
		||||
            if (consumer != null) {
 | 
			
		||||
                consumer.accept(new FromDeviceRpcResponse(requestId, null, RpcError.TIMEOUT));
 | 
			
		||||
@ -189,9 +189,9 @@ public class DefaultTbCoreDeviceRpcService implements TbCoreDeviceRpcService {
 | 
			
		||||
 | 
			
		||||
    private void scheduleToDeviceTimeout(ToDeviceRpcRequest request, UUID requestId) {
 | 
			
		||||
        long timeout = Math.max(0, request.getExpirationTime() - System.currentTimeMillis());
 | 
			
		||||
        log.trace("[{}] processing to device request: [{}]", this.hashCode(), requestId);
 | 
			
		||||
        log.trace("[{}] processing to device request.", requestId);
 | 
			
		||||
        scheduler.schedule(() -> {
 | 
			
		||||
            log.trace("[{}] timeout for to device request: [{}]", this.hashCode(), requestId);
 | 
			
		||||
            log.trace("[{}] timeout for to device request.", requestId);
 | 
			
		||||
            localToDeviceRpcRequests.remove(requestId);
 | 
			
		||||
        }, timeout, TimeUnit.MILLISECONDS);
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@ -622,7 +622,7 @@ queue:
 | 
			
		||||
          # For BATCH only
 | 
			
		||||
          batch-size: "${TB_QUEUE_RE_MAIN_SUBMIT_STRATEGY_BATCH_SIZE:1000}" # Maximum number of messages in batch
 | 
			
		||||
        processing-strategy:
 | 
			
		||||
          type: "${TB_QUEUE_RE_MAIN_PROCESSING_STRATEGY_TYPE:RETRY_FAILED_AND_TIMED_OUT}" # SKIP_ALL_FAILURES, RETRY_ALL, RETRY_FAILED, RETRY_TIMED_OUT, RETRY_FAILED_AND_TIMED_OUT
 | 
			
		||||
          type: "${TB_QUEUE_RE_MAIN_PROCESSING_STRATEGY_TYPE:SKIP_ALL_FAILURES}" # SKIP_ALL_FAILURES, RETRY_ALL, RETRY_FAILED, RETRY_TIMED_OUT, RETRY_FAILED_AND_TIMED_OUT
 | 
			
		||||
          # For RETRY_ALL, RETRY_FAILED, RETRY_TIMED_OUT, RETRY_FAILED_AND_TIMED_OUT
 | 
			
		||||
          retries: "${TB_QUEUE_RE_MAIN_PROCESSING_STRATEGY_RETRIES:3}" # Number of retries, 0 is unlimited
 | 
			
		||||
          failure-percentage: "${TB_QUEUE_RE_MAIN_PROCESSING_STRATEGY_FAILURE_PERCENTAGE:0}" # Skip retry if failures or timeouts are less then X percentage of messages;
 | 
			
		||||
 | 
			
		||||
@ -37,6 +37,7 @@ import org.thingsboard.server.service.security.AccessValidator;
 | 
			
		||||
import java.util.Arrays;
 | 
			
		||||
import java.util.concurrent.CountDownLatch;
 | 
			
		||||
import java.util.concurrent.TimeUnit;
 | 
			
		||||
import java.util.concurrent.atomic.AtomicInteger;
 | 
			
		||||
 | 
			
		||||
import static org.junit.Assert.assertEquals;
 | 
			
		||||
import static org.junit.Assert.assertNotNull;
 | 
			
		||||
@ -55,6 +56,8 @@ public abstract class AbstractMqttServerSideRpcIntegrationTest extends AbstractC
 | 
			
		||||
    private User tenantAdmin;
 | 
			
		||||
    private Long asyncContextTimeoutToUseRpcPlugin;
 | 
			
		||||
 | 
			
		||||
    private static final AtomicInteger atomicInteger = new AtomicInteger(2);
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
    @Before
 | 
			
		||||
    public void beforeTest() throws Exception {
 | 
			
		||||
@ -70,7 +73,7 @@ public abstract class AbstractMqttServerSideRpcIntegrationTest extends AbstractC
 | 
			
		||||
        tenantAdmin = new User();
 | 
			
		||||
        tenantAdmin.setAuthority(Authority.TENANT_ADMIN);
 | 
			
		||||
        tenantAdmin.setTenantId(savedTenant.getId());
 | 
			
		||||
        tenantAdmin.setEmail("tenant2@thingsboard.org");
 | 
			
		||||
        tenantAdmin.setEmail("tenant" + atomicInteger.getAndIncrement() + "@thingsboard.org");
 | 
			
		||||
        tenantAdmin.setFirstName("Joe");
 | 
			
		||||
        tenantAdmin.setLastName("Downs");
 | 
			
		||||
 | 
			
		||||
@ -130,7 +133,7 @@ public abstract class AbstractMqttServerSideRpcIntegrationTest extends AbstractC
 | 
			
		||||
        String accessToken = deviceCredentials.getCredentialsId();
 | 
			
		||||
        assertNotNull(accessToken);
 | 
			
		||||
 | 
			
		||||
        String setGpioRequest = "{\"method\":\"setGpio\",\"params\":{\"pin\": \"23\",\"value\": 1},\"timeout\": 6000}";
 | 
			
		||||
        String setGpioRequest = "{\"method\":\"setGpio\",\"params\":{\"pin\": \"24\",\"value\": 1},\"timeout\": 6000}";
 | 
			
		||||
        String deviceId = savedDevice.getId().getId().toString();
 | 
			
		||||
 | 
			
		||||
        doPostAsync("/api/plugins/rpc/oneway/" + deviceId, setGpioRequest, String.class, status().isRequestTimeout(),
 | 
			
		||||
@ -139,7 +142,7 @@ public abstract class AbstractMqttServerSideRpcIntegrationTest extends AbstractC
 | 
			
		||||
 | 
			
		||||
    @Test
 | 
			
		||||
    public void testServerMqttOneWayRpcDeviceDoesNotExist() throws Exception {
 | 
			
		||||
        String setGpioRequest = "{\"method\":\"setGpio\",\"params\":{\"pin\": \"23\",\"value\": 1}}";
 | 
			
		||||
        String setGpioRequest = "{\"method\":\"setGpio\",\"params\":{\"pin\": \"25\",\"value\": 1}}";
 | 
			
		||||
        String nonExistentDeviceId = UUIDs.timeBased().toString();
 | 
			
		||||
 | 
			
		||||
        String result = doPostAsync("/api/plugins/rpc/oneway/" + nonExistentDeviceId, setGpioRequest, String.class,
 | 
			
		||||
@ -169,7 +172,7 @@ public abstract class AbstractMqttServerSideRpcIntegrationTest extends AbstractC
 | 
			
		||||
 | 
			
		||||
        Thread.sleep(2000);
 | 
			
		||||
 | 
			
		||||
        String setGpioRequest = "{\"method\":\"setGpio\",\"params\":{\"pin\": \"23\",\"value\": 1}}";
 | 
			
		||||
        String setGpioRequest = "{\"method\":\"setGpio\",\"params\":{\"pin\": \"26\",\"value\": 1}}";
 | 
			
		||||
        String deviceId = savedDevice.getId().getId().toString();
 | 
			
		||||
 | 
			
		||||
        String result = doPostAsync("/api/plugins/rpc/twoway/" + deviceId, setGpioRequest, String.class, status().isOk());
 | 
			
		||||
@ -187,7 +190,7 @@ public abstract class AbstractMqttServerSideRpcIntegrationTest extends AbstractC
 | 
			
		||||
        String accessToken = deviceCredentials.getCredentialsId();
 | 
			
		||||
        assertNotNull(accessToken);
 | 
			
		||||
 | 
			
		||||
        String setGpioRequest = "{\"method\":\"setGpio\",\"params\":{\"pin\": \"23\",\"value\": 1},\"timeout\": 6000}";
 | 
			
		||||
        String setGpioRequest = "{\"method\":\"setGpio\",\"params\":{\"pin\": \"27\",\"value\": 1},\"timeout\": 6000}";
 | 
			
		||||
        String deviceId = savedDevice.getId().getId().toString();
 | 
			
		||||
 | 
			
		||||
        doPostAsync("/api/plugins/rpc/twoway/" + deviceId, setGpioRequest, String.class, status().isRequestTimeout(),
 | 
			
		||||
@ -196,7 +199,7 @@ public abstract class AbstractMqttServerSideRpcIntegrationTest extends AbstractC
 | 
			
		||||
 | 
			
		||||
    @Test
 | 
			
		||||
    public void testServerMqttTwoWayRpcDeviceDoesNotExist() throws Exception {
 | 
			
		||||
        String setGpioRequest = "{\"method\":\"setGpio\",\"params\":{\"pin\": \"23\",\"value\": 1}}";
 | 
			
		||||
        String setGpioRequest = "{\"method\":\"setGpio\",\"params\":{\"pin\": \"28\",\"value\": 1}}";
 | 
			
		||||
        String nonExistentDeviceId = UUIDs.timeBased().toString();
 | 
			
		||||
 | 
			
		||||
        String result = doPostAsync("/api/plugins/rpc/twoway/" + nonExistentDeviceId, setGpioRequest, String.class,
 | 
			
		||||
 | 
			
		||||
@ -107,13 +107,13 @@ public abstract class AbstractMqttTelemetryIntegrationTest extends AbstractContr
 | 
			
		||||
        CountDownLatch latch = new CountDownLatch(1);
 | 
			
		||||
        TestMqttCallback callback = new TestMqttCallback(client, latch);
 | 
			
		||||
        client.setCallback(callback);
 | 
			
		||||
        client.connect(options).waitForCompletion(3000);
 | 
			
		||||
        client.connect(options).waitForCompletion(5000);
 | 
			
		||||
        client.subscribe("v1/devices/me/attributes", MqttQoS.AT_MOST_ONCE.value());
 | 
			
		||||
        String payload = "{\"key\":\"value\"}";
 | 
			
		||||
//        TODO 3.1: we need to acknowledge subscription only after it is processed by device actor and not when the message is pushed to queue.
 | 
			
		||||
//        MqttClient -> SUB REQUEST -> Transport -> Kafka -> Device Actor (subscribed)
 | 
			
		||||
//        MqttClient <- SUB_ACK <- Transport
 | 
			
		||||
        Thread.sleep(1000);
 | 
			
		||||
        Thread.sleep(5000);
 | 
			
		||||
        doPostAsync("/api/plugins/telemetry/" + savedDevice.getId() + "/SHARED_SCOPE", payload, String.class, status().isOk());
 | 
			
		||||
        latch.await(10, TimeUnit.SECONDS);
 | 
			
		||||
        assertEquals(payload, callback.getPayload());
 | 
			
		||||
 | 
			
		||||
@ -152,7 +152,7 @@ public abstract class AbstractRuleEngineFlowIntegrationTest extends AbstractRule
 | 
			
		||||
        QueueToRuleEngineMsg qMsg = new QueueToRuleEngineMsg(savedTenant.getId(), tbMsg, null, null);
 | 
			
		||||
        // Pushing Message to the system
 | 
			
		||||
        actorSystem.tell(qMsg, ActorRef.noSender());
 | 
			
		||||
        Mockito.verify(tbMsgCallback, Mockito.timeout(3000)).onSuccess();
 | 
			
		||||
        Mockito.verify(tbMsgCallback, Mockito.timeout(10000)).onSuccess();
 | 
			
		||||
 | 
			
		||||
        TimePageData<Event> eventsPage = getDebugEvents(savedTenant.getId(), ruleChain.getFirstRuleNodeId(), 1000);
 | 
			
		||||
        List<Event> events = eventsPage.getData().stream().filter(filterByCustomEvent()).collect(Collectors.toList());
 | 
			
		||||
@ -265,7 +265,7 @@ public abstract class AbstractRuleEngineFlowIntegrationTest extends AbstractRule
 | 
			
		||||
        // Pushing Message to the system
 | 
			
		||||
        actorSystem.tell(qMsg, ActorRef.noSender());
 | 
			
		||||
 | 
			
		||||
        Mockito.verify(tbMsgCallback, Mockito.timeout(3000)).onSuccess();
 | 
			
		||||
        Mockito.verify(tbMsgCallback, Mockito.timeout(10000)).onSuccess();
 | 
			
		||||
 | 
			
		||||
        TimePageData<Event> eventsPage = getDebugEvents(savedTenant.getId(), rootRuleChain.getFirstRuleNodeId(), 1000);
 | 
			
		||||
        List<Event> events = eventsPage.getData().stream().filter(filterByCustomEvent()).collect(Collectors.toList());
 | 
			
		||||
 | 
			
		||||
@ -7,7 +7,7 @@
 | 
			
		||||
        </encoder>
 | 
			
		||||
    </appender>
 | 
			
		||||
 | 
			
		||||
    <logger name="org.thingsboard.server" level="WARN"/>
 | 
			
		||||
    <logger name="org.thingsboard.server" level="TRACE"/>
 | 
			
		||||
    <logger name="org.springframework" level="WARN"/>
 | 
			
		||||
    <logger name="org.springframework.boot.test" level="WARN"/>
 | 
			
		||||
    <logger name="org.apache.cassandra" level="WARN"/>
 | 
			
		||||
 | 
			
		||||
@ -50,10 +50,10 @@ public final class InMemoryStorage {
 | 
			
		||||
        return storage.computeIfAbsent(topic, (t) -> new LinkedBlockingQueue<>()).add(msg);
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    public <T extends TbQueueMsg> List<T> get(String topic, long durationInMillis) throws InterruptedException {
 | 
			
		||||
    public <T extends TbQueueMsg> List<T> get(String topic) throws InterruptedException {
 | 
			
		||||
        if (storage.containsKey(topic)) {
 | 
			
		||||
            List<T> entities;
 | 
			
		||||
            T first = (T) storage.get(topic).poll(durationInMillis, TimeUnit.MILLISECONDS);
 | 
			
		||||
            T first = (T) storage.get(topic).poll();
 | 
			
		||||
            if (first != null) {
 | 
			
		||||
                entities = new ArrayList<>();
 | 
			
		||||
                entities.add(first);
 | 
			
		||||
 | 
			
		||||
@ -68,7 +68,7 @@ public class InMemoryTbQueueConsumer<T extends TbQueueMsg> implements TbQueueCon
 | 
			
		||||
                    .stream()
 | 
			
		||||
                    .map(tpi -> {
 | 
			
		||||
                        try {
 | 
			
		||||
                            return storage.get(tpi.getFullTopicName(), durationInMillis);
 | 
			
		||||
                            return storage.get(tpi.getFullTopicName());
 | 
			
		||||
                        } catch (InterruptedException e) {
 | 
			
		||||
                            if (!stopped) {
 | 
			
		||||
                                log.error("Queue was interrupted.", e);
 | 
			
		||||
 | 
			
		||||
@ -24,7 +24,7 @@ public class TbRuleEngineQueueConfiguration {
 | 
			
		||||
    private String topic;
 | 
			
		||||
    private int pollInterval;
 | 
			
		||||
    private int partitions;
 | 
			
		||||
    private String packProcessingTimeout;
 | 
			
		||||
    private long packProcessingTimeout;
 | 
			
		||||
    private TbRuleEngineQueueSubmitStrategyConfiguration submitStrategy;
 | 
			
		||||
    private TbRuleEngineQueueAckStrategyConfiguration processingStrategy;
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@ -35,4 +35,12 @@ service.type=monolith
 | 
			
		||||
#spring.datasource.password=postgres
 | 
			
		||||
#spring.datasource.url=jdbc:postgresql://localhost:5432/sqltest
 | 
			
		||||
#spring.datasource.driverClassName=org.postgresql.Driver
 | 
			
		||||
#spring.datasource.hikari.maximumPoolSize = 50
 | 
			
		||||
#spring.datasource.hikari.maximumPoolSize = 50
 | 
			
		||||
 | 
			
		||||
queue.rule-engine.queues[0].name=Main
 | 
			
		||||
queue.rule-engine.queues[0].topic=tb_rule_engine.main
 | 
			
		||||
queue.rule-engine.queues[0].poll-interval=25
 | 
			
		||||
queue.rule-engine.queues[0].partitions=3
 | 
			
		||||
queue.rule-engine.queues[0].pack-processing-timeout=3000
 | 
			
		||||
queue.rule-engine.queues[0].processing-strategy.type=SKIP_ALL_FAILURES
 | 
			
		||||
queue.rule-engine.queues[0].submit-strategy.type=BURST
 | 
			
		||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user