Started by user Fabien Désiles Obtained tools/jenkins/qa/component-api-test/Jenkinsfile from git https://github.com/Talend/data-processing-runtime.git Loading library dp-runtime@main Attempting to resolve main from remote references... > git --version # timeout=10 > git --version # 'git version 2.30.2' using GIT_SSH to set credentials Verifying host key using known hosts file, will automatically accept unseen keys > git ls-remote -- git@github.com:Talend/data-processing-runtime-jenkins-shared-lib.git # timeout=10 Found match: refs/heads/main revision 8c2fd2a26690aca9b7778383bbe624e09e2d0e45 The recommended git tool is: NONE using credential github-ssh > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/QA/component-api-test-new@libs/846c20a2701585142ba3b75bd8453ccb919f657be3c78ac277767685aff111c2/.git # timeout=10 Fetching changes from the remote Git repository > git config remote.origin.url git@github.com:Talend/data-processing-runtime-jenkins-shared-lib.git # timeout=10 Fetching without tags Fetching upstream changes from git@github.com:Talend/data-processing-runtime-jenkins-shared-lib.git > git --version # timeout=10 > git --version # 'git version 2.30.2' using GIT_SSH to set credentials Verifying host key using known hosts file, will automatically accept unseen keys > git fetch --no-tags --force --progress -- git@github.com:Talend/data-processing-runtime-jenkins-shared-lib.git +refs/heads/*:refs/remotes/origin/* # timeout=10 Checking out Revision 8c2fd2a26690aca9b7778383bbe624e09e2d0e45 (main) > git config core.sparsecheckout # timeout=10 > git checkout -f 8c2fd2a26690aca9b7778383bbe624e09e2d0e45 # timeout=10 Commit message: "qa(TPRUN-4606): Component tests - dedicated tenants (#2)" > git rev-list --no-walk 8c2fd2a26690aca9b7778383bbe624e09e2d0e45 # timeout=10 [Pipeline] Start of Pipeline [Pipeline] echo [WARNING] label option is deprecated. To use a static pod template, use the 'inheritFrom' option. [Pipeline] podTemplate [Pipeline] { [Pipeline] node Created Pod: kubernetes jenkins-dp-runtime/data-processing-runtime-qa-component-test-789bc7dd-79-z36-1h2s4 Agent data-processing-runtime-qa-component-test-789bc7dd-79-z36-1h2s4 is provisioned from template data-processing-runtime-qa-component-test-789bc7dd-79-z368x --- apiVersion: "v1" kind: "Pod" metadata: annotations: buildUrl: "http://jenkins-dp-runtime.ci.svc.cluster.local:8080/job/QA/job/component-api-test-new/50/" runUrl: "job/QA/job/component-api-test-new/50/" labels: jenkins/jenkins-dp-runtime-jenkins-agent: "true" jenkins/label-digest: "ce8cd5b8e1ac1a39336ca139b7036d80937e27c2" jenkins/label: "data-processing-runtime-qa-component-test-789bc7dd-79" name: "data-processing-runtime-qa-component-test-789bc7dd-79-z36-1h2s4" namespace: "jenkins-dp-runtime" spec: containers: - command: - "cat" image: "artifactory.datapwn.com/tlnd-docker-prod/talend/common/tsbi/custom-builder:3.0.9-20220930152032" name: "custom-builder" resources: limits: memory: "2G" cpu: "1.5" requests: memory: "1G" cpu: "1" tty: true volumeMounts: - mountPath: "/root/.m2" name: "m2-cache" - mountPath: "/root/.asdf/installs" name: "efs-jenkins-dp-runtime-asdf" subPath: "installs" - mountPath: "/root/.asdf/shims" name: "efs-jenkins-dp-runtime-asdf" subPath: "shims" - mountPath: "/home/jenkins/agent" name: "workspace-volume" readOnly: false - env: - name: "JENKINS_SECRET" value: "********" - name: "JENKINS_TUNNEL" value: "jenkins-dp-runtime-agent.ci.svc.cluster.local:50000" - name: "JENKINS_AGENT_NAME" value: "data-processing-runtime-qa-component-test-789bc7dd-79-z36-1h2s4" - name: "JENKINS_NAME" value: "data-processing-runtime-qa-component-test-789bc7dd-79-z36-1h2s4" - name: "JENKINS_AGENT_WORKDIR" value: "/home/jenkins/agent" - name: "JENKINS_URL" value: "http://jenkins-dp-runtime.ci.svc.cluster.local:8080/" image: "jenkins/inbound-agent:4.11-1-jdk11" name: "jnlp" resources: limits: {} requests: memory: "256Mi" cpu: "100m" volumeMounts: - mountPath: "/home/jenkins/agent" name: "workspace-volume" readOnly: false imagePullSecrets: - name: "talend-registry" nodeSelector: kubernetes.io/os: "linux" restartPolicy: "Never" volumes: - name: "m2-cache" persistentVolumeClaim: claimName: "efs-jenkins-dp-runtime-m2" - emptyDir: medium: "" name: "workspace-volume" - name: "efs-jenkins-dp-runtime-asdf" persistentVolumeClaim: claimName: "efs-jenkins-dp-runtime-asdf" Running on data-processing-runtime-qa-component-test-789bc7dd-79-z36-1h2s4 in /home/jenkins/agent/workspace/QA/component-api-test-new [Pipeline] { [Pipeline] container [Pipeline] { [Pipeline] withEnv [Pipeline] { [Pipeline] ansiColor [Pipeline] { [Pipeline] timeout Timeout set to expire in 12 hr [Pipeline] { [Pipeline] stage [Pipeline] { (Checkout code) [Pipeline] dir Running in /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime [Pipeline] { [Pipeline] script [Pipeline] { [Pipeline] git The recommended git tool is: git using credential github-credentials Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@525c561c; decorates RemoteLauncher[hudson.remoting.Channel@3aa963b5:JNLP4-connect connection from ip-10-90-41-232.ec2.internal/10.90.41.232:41264] will be ignored (a typical symptom is the Git executable not being run inside a designated container) Cloning the remote Git repository Cloning repository https://github.com/Talend/data-processing-runtime.git > git init /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime # timeout=10 Fetching upstream changes from https://github.com/Talend/data-processing-runtime.git > git --version # timeout=10 > git --version # 'git version 2.30.2' using GIT_ASKPASS to set credentials > git fetch --tags --force --progress -- https://github.com/Talend/data-processing-runtime.git +refs/heads/*:refs/remotes/origin/* # timeout=10 Avoid second fetch Checking out Revision 1ce094b2c84807d608562793cf5ea438d2167167 (refs/remotes/origin/main) > git config remote.origin.url https://github.com/Talend/data-processing-runtime.git # timeout=10 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10 > git rev-parse refs/remotes/origin/main^{commit} # timeout=10 > git config core.sparsecheckout # timeout=10 > git checkout -f 1ce094b2c84807d608562793cf5ea438d2167167 # timeout=10 > git branch -a -v --no-abbrev # timeout=10 > git checkout -b main 1ce094b2c84807d608562793cf5ea438d2167167 # timeout=10 Commit message: "fix(TPRUN-4562): R2022-11 CVEs (#920)" > git rev-list --no-walk c157e06a70de83aeb0cdc85cd5f14b9ba1ded891 # timeout=10 [Pipeline] } [Pipeline] // script [Pipeline] } [Pipeline] // dir [Pipeline] } [Pipeline] // stage [Pipeline] stage [Pipeline] { (Set up worker) [Pipeline] script [Pipeline] { [Pipeline] echo Install apps in pod [Pipeline] sh + bash /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/tools/jenkins/qa/component-api-test/scripts/asdf-install.sh maven 3.6.0 is already installed jq 1.5 is already installed java openjdk-11 is already installed [Pipeline] } [Pipeline] // script [Pipeline] } [Pipeline] // stage [Pipeline] stage [Pipeline] { (Set up job configuration) [Pipeline] script [Pipeline] { [Pipeline] withCredentials Masking supported pattern matches of $tpdQaDevPassword or $tpdTestEnvpassword or $tpdQaProdPassword [Pipeline] { [Pipeline] echo Setting up stack configuration properties [Pipeline] libraryResource [Pipeline] writeFile [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k testEnvUrl -v https://tpd.at.cloud.talend.com -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/stack-configuration.properties [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k accountName -v tdi-com-test.talend.com -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/stack-configuration.properties [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k user -v tdi-com-test -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/stack-configuration.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [tpdQaDevPassword] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k password -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/stack-configuration.properties [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k remoteEngineId -v 9e046f3d-c9e5-406b-bdf6-309599d628d4 -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/stack-configuration.properties [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k profileId -v 63722a73345624388af6f683 -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/stack-configuration.properties [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k tmcBaseUrl -v https://tmc.at.cloud.talend.com -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/stack-configuration.properties [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k tmcCe4dUrl -v https://tmc.at.cloud.talend.com/api/ipaas/services/admin/remote/elastic-engines/ -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/stack-configuration.properties [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k tmcCe4dId -v '' -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/stack-configuration.properties [Pipeline] } [Pipeline] // withCredentials [Pipeline] } [Pipeline] // script [Pipeline] } [Pipeline] // stage [Pipeline] stage [Pipeline] { (Set up test environment) [Pipeline] script [Pipeline] { [Pipeline] withCredentials Masking supported pattern matches of $deltaLakeToken [Pipeline] { [Pipeline] libraryResource [Pipeline] writeFile [Pipeline] echo Start databricks cluster (if not already running) [Pipeline] sh + timeout 5m bash ./.jenkins_shared_lib_scripts/qa/scripts/start_databricks_cluster.sh **** Check current databricks cluster state Cluster state is RUNNING : Cluster already started [Pipeline] } [Pipeline] // withCredentials [Pipeline] withCredentials Masking supported pattern matches of $nexusPassword [Pipeline] { [Pipeline] libraryResource [Pipeline] writeFile [Pipeline] echo Set Maven configuration file [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_maven_settings_file.sh -f /root/.m2/settings.xml -u ci-tpd -p **** [Pipeline] } [Pipeline] // withCredentials [Pipeline] withCredentials Masking supported pattern matches of $nexusPassword or $googleProjectId or $workdayRaaSPwd or $workdayRaaSTenantAlias or $snowflakePrivateKeyPwd or $snowflakePrivateKey or $dynamoAssumeRoleAccessKey or $dynamoAssumeRoleSecretKey or $cosmosKey or $accessKey or $secretKey or $eventHubsKey or $netSuitePwd or $netSuiteAppID or $azureDynamics365Pwd or $azureDynamics365AppId or $googlePrivateId or $googlePrivateKey or $workdayClientId or $workdayClientSecret or $postgresqlPwd or $mariadbPwd or $mysqlPwd or $auroraPwd or $mssqlPwd or $redShiftPwd or $snowflakePwd or $azureSQLDWHPwd or $azureBlobAccountKey or $azureDataLakeSharedKey or $salesforcePwd or $marketoClientSecret or $salesforceSecretKey or $sftpPassword or $ftpPassword or $mongodbPassword or $singlestorePassword or $derbyPassword or $couchbasePassword or $azureDatalakeSas or $azureBlobSas or $deltaLakeToken or $boxClientSecret or $boxKeyPassphrase or $boxPrivateKey [Pipeline] { [Pipeline] echo Setting up test environment configuration properties [Pipeline] echo File : amazon-parameters.properties [Pipeline] libraryResource [Pipeline] writeFile [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [accessKey] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k accessKey -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/amazon-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [secretKey] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k secretKey -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/amazon-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [dynamoAssumeRoleAccessKey] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k dynamoAssumeRoleAccessKey -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/amazon-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [dynamoAssumeRoleSecretKey] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k dynamoAssumeRoleSecretKey -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/amazon-parameters.properties [Pipeline] echo File : azure-parameters.properties [Pipeline] libraryResource [Pipeline] writeFile [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [azureDatalakeSas] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k azureDatalakeSas -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/azure-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [azureBlobSas] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k azureBlobSas -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/azure-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [azureBlobAccountKey] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k azureBlobAccountKey -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/azure-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [azureDataLakeSharedKey] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k azureDataLakeSharedKey -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/azure-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [azureDynamics365Pwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k azureDynamics365Pwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/azure-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [azureDynamics365AppId] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k azureDynamics365AppId -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/azure-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [eventHubsKey] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k eventHubsKey -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/azure-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [cosmosKey] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k cosmosKey -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/azure-parameters.properties [Pipeline] echo File : database-parameters.properties [Pipeline] libraryResource [Pipeline] writeFile [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [postgresqlPwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k postgresqlPwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [mariadbPwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k mariadbPwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [mysqlPwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k mysqlPwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [auroraPwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k auroraPwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [mssqlPwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k mssqlPwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [redShiftPwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k redShiftPwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [deltaLakeToken] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k deltaLakeToken -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [snowflakePwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k snowflakePwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [snowflakePrivateKeyPwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k snowflakePrivateKeyPwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [snowflakePrivateKey] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k snowflakePrivateKey -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [azureSQLDWHPwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k azureSQLDWHPwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k singlestoreUsername -v root -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [singlestorePassword] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k singlestorePassword -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [derbyPassword] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k derbyUsername -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [derbyPassword] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k derbyPassword -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/database-parameters.properties [Pipeline] echo File : common-parameters.properties [Pipeline] libraryResource [Pipeline] writeFile [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k marketoClientId -v a1705bd3-9530-44d2-acce-d2d9fae9347c -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [marketoClientSecret] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k marketoClientSecret -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k sftpUsername -v sftp-dpp-runtime-user -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [sftpPassword] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k sftpPassword -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k ftpUsername -v ftpuser -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [mongodbPassword, ftpPassword] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k ftpPassword -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [postgresqlPwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k mongodbUsername -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [mongodbPassword, ftpPassword] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k mongodbPassword -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k couchbaseUsername -v couchbase-dpp-runtime-user -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [couchbasePassword] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k couchbasePassword -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [salesforcePwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k salesforcePwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [salesforceSecretKey] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k salesforceSecretKey -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [workdayClientId] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k workdayClientId -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [workdayClientSecret] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k workdayClientSecret -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [workdayRaaSPwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k workdayRaaSPwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [workdayRaaSTenantAlias, postgresqlPwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k workdayRaaSTenantAlias -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [netSuitePwd, postgresqlPwd] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k netSuitePwd -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [netSuiteAppID] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k netSuiteAppID -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [boxClientSecret] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k boxClientSecret -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [boxKeyPassphrase] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k boxKeyPassphrase -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [boxPrivateKey] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k boxPrivateKey -v '****' -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/common-parameters.properties [Pipeline] echo File : google-parameters.properties [Pipeline] libraryResource [Pipeline] writeFile [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [googlePrivateId] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k googlePrivateId -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/google-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [googlePrivateKey] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k googlePrivateKey -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/google-parameters.properties [Pipeline] sh Warning: A secret was passed to "sh" using Groovy String interpolation, which is insecure. Affected argument(s) used the following variable(s): [googleProjectId] See https://jenkins.io/redirect/groovy-string-interpolation for details. + bash ./.jenkins_shared_lib_scripts/qa/scripts/set_properties.sh -k googleProjectId -v **** -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/resources/properties/google-parameters.properties [Pipeline] } [Pipeline] // withCredentials [Pipeline] } [Pipeline] // script [Pipeline] } [Pipeline] // stage [Pipeline] stage [Pipeline] { (Clean up before test) [Pipeline] script [Pipeline] { [Pipeline] sh + echo 'Remove all existing datasets/connectors/pipelines/datapreps' Remove all existing datasets/connectors/pipelines/datapreps + bash mvn -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/pom.xml test '-Dcucumber.options=--tags '\''@delete'\''' -Dauth0=true [WARNING] [WARNING] Some problems were encountered while building the effective settings [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] [INFO] Scanning for projects... [INFO] [INFO] ---------< org.talend.qa.test:dpruntime-tests-api-components >---------- [INFO] Building dpruntime-tests-api-components 1.0.0-SNAPSHOT [INFO] --------------------------------[ jar ]--------------------------------- [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ dpruntime-tests-api-components --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] skip non existing resourceDirectory /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/main/resources [INFO] [INFO] --- maven-compiler-plugin:3.7.0:compile (default-compile) @ dpruntime-tests-api-components --- [INFO] No sources to compile [INFO] [INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ dpruntime-tests-api-components --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 352 resources [INFO] [INFO] --- maven-compiler-plugin:3.7.0:testCompile (default-testCompile) @ dpruntime-tests-api-components --- [INFO] Changes detected - recompiling the module! [INFO] Compiling 23 source files to /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/target/test-classes [INFO] /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/java/utils/TPDAPIRequest.java: Some input files use or override a deprecated API. [INFO] /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/test/java/utils/TPDAPIRequest.java: Recompile with -Xlint:deprecation for details. [INFO] [INFO] --- maven-surefire-plugin:2.12.4:test (default-test) @ dpruntime-tests-api-components --- [INFO] Surefire report directory: /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/target/surefire-reports ------------------------------------------------------- T E S T S ------------------------------------------------------- Running CucumberTest @delete Feature: Delete Connections , Datasets and Pipelines @delete @delete Scenario: delete resources # features/preparation/DeleteResources.feature:4 WARNING: An illegal reflective access operation has occurred WARNING: Illegal reflective access by cucumber.deps.com.thoughtworks.xstream.core.util.Fields (file:/root/.m2/repository/io/cucumber/cucumber-jvm-deps/1.0.6/cucumber-jvm-deps-1.0.6.jar) to field java.util.TreeMap.comparator WARNING: Please consider reporting this to the maintainers of cucumber.deps.com.thoughtworks.xstream.core.util.Fields WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations WARNING: All illegal access operations will be denied in a future release SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". SLF4J: Defaulting to no-operation (NOP) logger implementation SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details. When user cleans up pipelines after run # UtilStep.cleanUp(String) When user cleans up preparatons after run # UtilStep.cleanUp(String) When user cleans up datasets after run # UtilStep.cleanUp(String) When user cleans up connections after run # UtilStep.cleanUp(String) 1 Scenarios (1 passed) 4 Steps (4 passed) 23m54.325s Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1,435.185 sec Results : Tests run: 1, Failures: 0, Errors: 0, Skipped: 0 [INFO] ------------------------------------------------------------------------ [INFO] BUILD SUCCESS [INFO] ------------------------------------------------------------------------ [INFO] Total time: 24:11 min [INFO] Finished at: 2022-11-14T13:48:40Z [INFO] ------------------------------------------------------------------------ [Pipeline] } [Pipeline] // script [Pipeline] } [Pipeline] // stage [Pipeline] stage [Pipeline] { (Prepare) [Pipeline] script [Pipeline] { [Pipeline] echo the connection tag is @component-Cnt [Pipeline] echo the final connection tag is @component-Cnt or @saveCntId or @Test-Cnt [Pipeline] sh + bash mvn test -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/pom.xml '-Dcucumber.options=--tags '\''@component-Cnt or @saveCntId or @Test-Cnt'\''' -Dauth0=true [WARNING] [WARNING] Some problems were encountered while building the effective settings [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] [INFO] Scanning for projects... [INFO] [INFO] ---------< org.talend.qa.test:dpruntime-tests-api-components >---------- [INFO] Building dpruntime-tests-api-components 1.0.0-SNAPSHOT [INFO] --------------------------------[ jar ]--------------------------------- [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ dpruntime-tests-api-components --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] skip non existing resourceDirectory /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/main/resources [INFO] [INFO] --- maven-compiler-plugin:3.7.0:compile (default-compile) @ dpruntime-tests-api-components --- [INFO] No sources to compile [INFO] [INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ dpruntime-tests-api-components --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 352 resources [INFO] [INFO] --- maven-compiler-plugin:3.7.0:testCompile (default-testCompile) @ dpruntime-tests-api-components --- [INFO] Nothing to compile - all classes are up to date [INFO] [INFO] --- maven-surefire-plugin:2.12.4:test (default-test) @ dpruntime-tests-api-components --- [INFO] Surefire report directory: /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/target/surefire-reports ------------------------------------------------------- T E S T S ------------------------------------------------------- Running CucumberTest @sample-Cnt Feature: inject Amazon connection @sample-Cnt @component-Cnt @AmazonKinesis-Cnt @smoke-Cnt Scenario: inject KinesisSTS connection # features/preparation/connections/AmazonConnectionPreparation.feature:5 WARNING: An illegal reflective access operation has occurred WARNING: Illegal reflective access by cucumber.deps.com.thoughtworks.xstream.core.util.Fields (file:/root/.m2/repository/io/cucumber/cucumber-jvm-deps/1.0.6/cucumber-jvm-deps-1.0.6.jar) to field java.util.TreeMap.comparator WARNING: Please consider reporting this to the maintainers of cucumber.deps.com.thoughtworks.xstream.core.util.Fields WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations WARNING: All illegal access operations will be denied in a future release SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". SLF4J: Defaulting to no-operation (NOP) logger implementation SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details. Nov 14, 2022 1:49:11 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection AmazonKinesisSTS is : 201 When user adds none-database AmazonKinesisSTS connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @smoke-Cnt @AmazonKinesis-Cnt Scenario: inject Kinesis connection # features/preparation/connections/AmazonConnectionPreparation.feature:10 Nov 14, 2022 1:49:14 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection AmazonKinesis is : 201 When user adds none-database AmazonKinesis connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @AmazonDynamoDB-Cnt @smoke-Cnt Scenario: inject DynamoDB connection # features/preparation/connections/AmazonConnectionPreparation.feature:14 Nov 14, 2022 1:49:16 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection AmazonDynamoDB is : 201 When user adds none-database AmazonDynamoDB connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @AmazonDynamoDB_AssumeRole-Cnt @smoke-Cnt Scenario: inject DynamoDB_AssumeRole connection # features/preparation/connections/AmazonConnectionPreparation.feature:18 Nov 14, 2022 1:49:19 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection AmazonDynamoDB_AssumeRole is : 201 When user adds none-database AmazonDynamoDB_AssumeRole connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @AmazonS3-Cnt @smoke-Cnt Scenario: inject S3 connection # features/preparation/connections/AmazonConnectionPreparation.feature:22 Nov 14, 2022 1:49:21 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection AmazonS3 is : 201 When user adds none-database AmazonS3 connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt Feature: inject Azure connections @sample-Cnt @component-Cnt @AzureDynamics365-Cnt Scenario: inject AzureDynamics365 connection # features/preparation/connections/AzureConnectionPreparation.feature:5 Nov 14, 2022 1:49:23 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection AzureDynamics365 is : 201 When user adds none-database AzureDynamics365 connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @AzureEventHubs-Cnt @smoke-Cnt Scenario: inject AzureEventHubs connection # features/preparation/connections/AzureConnectionPreparation.feature:9 Nov 14, 2022 1:49:25 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection AzureEventHubs is : 201 When user adds none-database AzureEventHubs connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @AzureCosmosDB-Cnt @smoke-Cnt Scenario: inject AzureCosmosDB connection # features/preparation/connections/AzureConnectionPreparation.feature:13 Nov 14, 2022 1:49:27 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection AzureCosmosDB is : 201 When user adds none-database AzureCosmosDB connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @AzureDataLakeStorageGen2-Cnt @smoke-Cnt Scenario: inject AzureDataLakeStorageGen2 connection # features/preparation/connections/AzureConnectionPreparation.feature:17 Nov 14, 2022 1:49:29 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection AzureDataLakeStorageGen2 is : 201 When user adds none-database AzureDataLakeStorageGen2 connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @smoke-Cnt @component-Cnt @AzureDataLakeStorageGen2SAS-Cnt Scenario: inject AzureDataLakeStorageGen2SAS connection # features/preparation/connections/AzureConnectionPreparation.feature:21 Nov 14, 2022 1:49:32 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection AzureDataLakeStorageGen2SAS is : 201 When user adds none-database AzureDataLakeStorageGen2SAS connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @AzureBlobStorage-Cnt @smoke-Cnt @AzureEventHubs-Cnt Scenario: inject AzureBlobStorage connection # features/preparation/connections/AzureConnectionPreparation.feature:29 Nov 14, 2022 1:49:34 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection AzureBlobStorage is : 201 When user adds none-database AzureBlobStorage connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @smoke-Cnt @component-Cnt @AzureBlobStorageSAS-Cnt @AzureEventHubs-Cnt Scenario: inject AzureBlobStorageSAS connection # features/preparation/connections/AzureConnectionPreparation.feature:37 Nov 14, 2022 1:49:36 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection AzureBlobStorageSAS is : 201 When user adds none-database AzureBlobStorageSAS connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt Feature: inject common services connection @sample-Cnt @component-Cnt @Test-Cnt Scenario: inject test connection # features/preparation/connections/CommonConnectionPreparation.feature:5 Nov 14, 2022 1:49:38 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection Test is : 201 When user adds none-database Test connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @ElasticSearch-Cnt @smoke-Cnt Scenario: inject ElasticSearch connection # features/preparation/connections/CommonConnectionPreparation.feature:9 Nov 14, 2022 1:49:40 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection ElasticSearch is : 201 When user adds none-database ElasticSearch connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @Kafka-Cnt @smoke-Cnt Scenario: inject Kafka connection # features/preparation/connections/CommonConnectionPreparation.feature:13 Nov 14, 2022 1:49:42 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection Kafka is : 201 When user adds none-database Kafka connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @REST-Cnt @smoke-Cnt @SourceIsTarget-Cnt Scenario: inject REST connection # features/preparation/connections/CommonConnectionPreparation.feature:17 Nov 14, 2022 1:49:44 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection REST is : 201 When user adds none-database REST connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @REST-Cnt @smoke-Cnt Scenario: inject REST_HTTPBin connection # features/preparation/connections/CommonConnectionPreparation.feature:21 Nov 14, 2022 1:49:46 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection REST_HTTPBin is : 201 When user adds none-database REST_HTTPBin connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @REST-Cnt @smoke-Cnt Scenario: inject REST_Authentication_Basic connection # features/preparation/connections/CommonConnectionPreparation.feature:25 Nov 14, 2022 1:49:49 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection REST_Authentication_Basic is : 201 When user adds none-database REST_Authentication_Basic connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @REST-Cnt @smoke-Cnt Scenario: inject REST_Authentication_Digest connection # features/preparation/connections/CommonConnectionPreparation.feature:29 Nov 14, 2022 1:49:51 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection REST_Authentication_Digest is : 201 When user adds none-database REST_Authentication_Digest connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @MongoDB-Cnt @smoke-Cnt Scenario: inject MongoDB connection # features/preparation/connections/CommonConnectionPreparation.feature:33 Nov 14, 2022 1:49:53 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection MongoDB is : 201 When user adds none-database MongoDB connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @Pulsar-Cnt @smoke-Cnt Scenario: inject Pulsar connection # features/preparation/connections/CommonConnectionPreparation.feature:37 Nov 14, 2022 1:49:55 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection Pulsar is : 201 When user adds none-database Pulsar connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @Marketo-Cnt @smoke-Cnt Scenario: inject Marketo connection # features/preparation/connections/CommonConnectionPreparation.feature:41 Nov 14, 2022 1:49:57 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection Marketo is : 201 When user adds none-database Marketo connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @Salesforce-Cnt @smoke-Cnt Scenario: inject Salesforce connection # features/preparation/connections/CommonConnectionPreparation.feature:45 Nov 14, 2022 1:49:59 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection Salesforce is : 201 When user adds none-database Salesforce connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @Workday-Cnt @smoke-Cnt Scenario: inject Workday connection with RaaS # features/preparation/connections/CommonConnectionPreparation.feature:52 Nov 14, 2022 1:50:01 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection WorkdayRaaS is : 201 When user adds none-database WorkdayRaaS connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @Workday-Cnt @smoke-Cnt Scenario: inject Workday connection with WSSA # features/preparation/connections/CommonConnectionPreparation.feature:56 Nov 14, 2022 1:50:03 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection WorkdayWSSA is : 201 When user adds none-database WorkdayWSSA connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @NetSuite-Cnt @smoke-Cnt Scenario: inject NetSuite connection # features/preparation/connections/CommonConnectionPreparation.feature:60 Nov 14, 2022 1:50:05 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection NetSuite is : 201 When user adds none-database NetSuite connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @HDFS-Cnt @smoke-Cnt Scenario: inject HDFS connection # features/preparation/connections/CommonConnectionPreparation.feature:64 Nov 14, 2022 1:50:08 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection HDFS is : 201 When user adds none-database HDFS connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @FTP-Cnt @smoke-Cnt Scenario: inject FTP connection # features/preparation/connections/CommonConnectionPreparation.feature:68 Nov 14, 2022 1:50:10 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection FTP is : 201 When user adds none-database FTP connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @SFTP-Cnt @smoke-Cnt Scenario: inject SFTP connection # features/preparation/connections/CommonConnectionPreparation.feature:72 Nov 14, 2022 1:50:12 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection SFTP is : 201 When user adds none-database SFTP connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @Zendesk-Cnt @smoke-Cnt Scenario: inject Zendesk connection # features/preparation/connections/CommonConnectionPreparation.feature:76 Nov 14, 2022 1:50:14 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection Zendesk is : 201 When user adds none-database Zendesk connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @LocalConnection-Cnt @smoke-Cnt Scenario: inject Local Connection # features/preparation/connections/CommonConnectionPreparation.feature:80 Nov 14, 2022 1:50:16 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection LocalConnection is : 201 When user adds none-database LocalConnection connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @Couchbase-Cnt @smoke-Cnt Scenario: inject Couchbase connection # features/preparation/connections/CommonConnectionPreparation.feature:84 Nov 14, 2022 1:50:18 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection Couchbase is : 201 When user adds none-database Couchbase connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @Kudu-Cnt @smoke-Cnt Scenario: inject Kudu connection # features/preparation/connections/CommonConnectionPreparation.feature:88 Nov 14, 2022 1:50:20 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection Kudu is : 201 When user adds none-database Kudu connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @Box-Cnt @smoke-Cnt @SourceIsTarget Scenario: inject Box_Enterprise_Id Connection # features/preparation/connections/CommonConnectionPreparation.feature:92 Nov 14, 2022 1:50:20 PM utils.payloadAction.connection.BoxConnectionPayloadAction updateBoxConnectionPara INFO: The length of Box key: 1834 Nov 14, 2022 1:50:22 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection Box_Enterprise_Id is : 201 When user adds none-database Box_Enterprise_Id connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @Box-Cnt @smoke-Cnt @SourceIsTarget Scenario: inject Box_User_Id Connection # features/preparation/connections/CommonConnectionPreparation.feature:96 Nov 14, 2022 1:50:22 PM utils.payloadAction.connection.BoxConnectionPayloadAction updateBoxConnectionPara INFO: The length of Box key: 1834 Nov 14, 2022 1:50:25 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection Box_User_Id is : 201 When user adds none-database Box_User_Id connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @Database-Cnt Feature: inject Database connection @sample-Cnt @Database-Cnt @component-Cnt @PostgreSQL-Cnt @smoke-Cnt Scenario: inject PostgreSQL connection # features/preparation/connections/DatabaseConnectionPreparation.feature:5 Nov 14, 2022 1:50:27 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection PostgreSQL is : 201 When user adds database PostgreSQL connection with dbType PostgreSQL # AddConnectionStep.addDatabaseConnection(String,String) @sample-Cnt @Database-Cnt @component-Cnt @PostgreSQL-Cnt @smoke-Cnt Scenario: inject PostgreSQL connection with Basic authentication type # features/preparation/connections/DatabaseConnectionPreparation.feature:8 Nov 14, 2022 1:50:29 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection PostgreSQLBasic is : 201 When user adds database PostgreSQL connection using dbType PostgreSQL and authenticationType Basic # AddConnectionStep.addDatabaseConnectionWithAuthenticationType(String,String,String) @sample-Cnt @Database-Cnt @component-Cnt @Aurora-Cnt @smoke-Cnt Scenario: inject Aurora connection # features/preparation/connections/DatabaseConnectionPreparation.feature:12 Nov 14, 2022 1:50:31 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection Aurora is : 201 When user adds database Aurora connection using dbType Aurora and configuration handler=MySQL # AddConnectionStep.addDatabaseConnection(String,String,String) @sample-Cnt @Database-Cnt @component-Cnt @MariaDB-Cnt @smoke-Cnt Scenario: inject MariaDB connection # features/preparation/connections/DatabaseConnectionPreparation.feature:16 Nov 14, 2022 1:50:33 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection MariaDB is : 201 When user adds database MariaDB connection with dbType MariaDB # AddConnectionStep.addDatabaseConnection(String,String) @sample-Cnt @Database-Cnt @component-Cnt @MariaDB-Cnt @smoke-Cnt Scenario: inject MariaDB connection with Basic authentication type # features/preparation/connections/DatabaseConnectionPreparation.feature:20 Nov 14, 2022 1:50:35 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection MariaDBBasic is : 201 When user adds database MariaDB connection using dbType MariaDB and authenticationType Basic # AddConnectionStep.addDatabaseConnectionWithAuthenticationType(String,String,String) @sample-Cnt @Database-Cnt @component-Cnt @MSSQL-Cnt @smoke-Cnt Scenario: inject MSSQL connection # features/preparation/connections/DatabaseConnectionPreparation.feature:24 Nov 14, 2022 1:50:37 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection MSSQL is : 201 When user adds database MSSQL connection with dbType MSSQL # AddConnectionStep.addDatabaseConnection(String,String) @sample-Cnt @Database-Cnt @component-Cnt @MSSQL-Cnt @smoke-Cnt Scenario: inject MSSQL connection with Basic authentication type # features/preparation/connections/DatabaseConnectionPreparation.feature:28 Nov 14, 2022 1:50:39 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection MSSQLBasic is : 201 When user adds database MSSQL connection using dbType MSSQL and authenticationType Basic # AddConnectionStep.addDatabaseConnectionWithAuthenticationType(String,String,String) @sample-Cnt @Database-Cnt @component-Cnt @MySql-Cnt @smoke-Cnt Scenario: inject MySQL connection # features/preparation/connections/DatabaseConnectionPreparation.feature:32 Nov 14, 2022 1:50:41 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection MySQL is : 201 When user adds database MySQL connection with dbType MySQL # AddConnectionStep.addDatabaseConnection(String,String) @sample-Cnt @Database-Cnt @component-Cnt @MySql-Cnt @smoke-Cnt Scenario: inject MySQL connection with Basic authentication type # features/preparation/connections/DatabaseConnectionPreparation.feature:36 Nov 14, 2022 1:50:44 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection MySQLBasic is : 201 When user adds database MySQL connection using dbType MySQL and authenticationType Basic # AddConnectionStep.addDatabaseConnectionWithAuthenticationType(String,String,String) @sample-Cnt @Database-Cnt @component-Cnt @RedShift-Cnt @smoke-Cnt Scenario: inject Redshift connection # features/preparation/connections/DatabaseConnectionPreparation.feature:40 Nov 14, 2022 1:50:46 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection Redshift is : 201 When user adds database Redshift connection with dbType Redshift # AddConnectionStep.addDatabaseConnection(String,String) @sample-Cnt @Database-Cnt @component-Cnt @Snowflake-Cnt @smoke-Cnt Scenario: inject Snowflake with basic connection # features/preparation/connections/DatabaseConnectionPreparation.feature:44 Nov 14, 2022 1:50:48 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection Snowflake is : 201 When user adds database Snowflake connection using dbType Snowflake and configuration authenticationType=BASIC # AddConnectionStep.addDatabaseConnection(String,String,String) @sample-Cnt @Database-Cnt @component-Cnt @Snowflake_Keypair-Cnt Scenario: inject Snowflake with Keypair connection # features/preparation/connections/DatabaseConnectionPreparation.feature:48 Nov 14, 2022 1:50:50 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection Snowflake-Keypair is : 201 When user adds database Snowflake-Keypair connection using dbType Snowflake and configuration authenticationType=KEY_PAIR # AddConnectionStep.addDatabaseConnection(String,String,String) @sample-Cnt @Database-Cnt @component-Cnt @Snowflake-Cnt @smoke-Cnt Scenario: inject Snowflake Basic authentication with parameter connection # features/preparation/connections/DatabaseConnectionPreparation.feature:52 Nov 14, 2022 1:50:52 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection SnowflakeBasic is : 201 When user adds database Snowflake connection using dbType Snowflake with authenticationType Basic and parameters schema&warehouse # AddConnectionStep.addDatabaseConnectionWithAuthenticationType(String,String,String,String) @sample-Cnt @Database-Cnt @component-Cnt @AzureSynapse-Cnt @smoke-Cnt Scenario: inject AzureSynapse connection # features/preparation/connections/DatabaseConnectionPreparation.feature:56 Nov 14, 2022 1:50:54 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection AzureSynapse is : 201 When user adds database AzureSynapse connection with dbType SQLDWH # AddConnectionStep.addDatabaseConnection(String,String) @sample-Cnt @Database-Cnt @component-Cnt @Derby-Cnt @smoke-Cnt Scenario: inject Derby connection # features/preparation/connections/DatabaseConnectionPreparation.feature:60 Nov 14, 2022 1:50:56 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection Derby is : 201 When user adds database Derby connection with dbType Derby # AddConnectionStep.addDatabaseConnection(String,String) @sample-Cnt @Database-Cnt @component-Cnt @SingleStore-Cnt @smoke-Cnt Scenario: inject SingleStore connection # features/preparation/connections/DatabaseConnectionPreparation.feature:64 Nov 14, 2022 1:50:58 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection SingleStore is : 201 When user adds database SingleStore connection using dbType SingleStore and configuration handler=MariaDB # AddConnectionStep.addDatabaseConnection(String,String,String) @sample-Cnt @Database-Cnt @component-Cnt @DeltaLake-Cnt @smoke-Cnt Scenario: inject DeltaLake connection # features/preparation/connections/DatabaseConnectionPreparation.feature:68 Nov 14, 2022 1:51:00 PM actions.TPDAPIOperation addDatabaseConnectionWithAuthType INFO: Status code for adding connection DeltaLake is : 201 When user adds database DeltaLake connection with dbType DeltaLake # AddConnectionStep.addDatabaseConnection(String,String) @sample-Cnt Feature: inject Google connections @sample-Cnt @component-Cnt @GoogleBigQuery-Cnt @smoke-Cnt Scenario: inject GoogleBigQuery connection # features/preparation/connections/GoogleConnectionPreparation.feature:5 Nov 14, 2022 1:51:03 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection GoogleBigQuery is : 201 When user adds none-database GoogleBigQuery connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @GooglePubSub-Cnt @smoke-Cnt Scenario: inject GooglePubSub connection # features/preparation/connections/GoogleConnectionPreparation.feature:9 Nov 14, 2022 1:51:05 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection GooglePubSub is : 201 When user adds none-database GooglePubSub connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt @component-Cnt @GoogleCloudStorage-Cnt @smoke-Cnt Scenario: inject GoogleCloudStorage connection # features/preparation/connections/GoogleConnectionPreparation.feature:13 Nov 14, 2022 1:51:07 PM actions.TPDAPIOperation userAddsConnection INFO: Status code for adding connection GoogleCloudStorage is : 201 When user adds none-database GoogleCloudStorage connection # AddConnectionStep.userAddsConnection(String) @sample-Cnt Feature: save connections' Id @sample-Cnt @saveCntId Scenario: put all connections' id in one file # features/preparation/connections/WeSaveConnectionId.feature:5 When add a json file to contain all the connection id # UtilStep.saveConnectionsId() 56 Scenarios (56 passed) 56 Steps (56 passed) 2m3.630s Tests run: 56, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 124.422 sec Results : Tests run: 56, Failures: 0, Errors: 0, Skipped: 0 [INFO] ------------------------------------------------------------------------ [INFO] BUILD SUCCESS [INFO] ------------------------------------------------------------------------ [INFO] Total time: 02:13 min [INFO] Finished at: 2022-11-14T13:51:07Z [INFO] ------------------------------------------------------------------------ [Pipeline] echo the dataset tag is @component-ds [Pipeline] echo the final dataset tag is @component-ds or @saveDsId or @Test-ds [Pipeline] sh + bash mvn test -f /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/pom.xml '-Dcucumber.options=--tags '\''@component-ds or @saveDsId or @Test-ds'\''' -Dauth0=true [WARNING] [WARNING] Some problems were encountered while building the effective settings [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] [INFO] Scanning for projects... [INFO] [INFO] ---------< org.talend.qa.test:dpruntime-tests-api-components >---------- [INFO] Building dpruntime-tests-api-components 1.0.0-SNAPSHOT [INFO] --------------------------------[ jar ]--------------------------------- [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ dpruntime-tests-api-components --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] skip non existing resourceDirectory /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/main/resources [INFO] [INFO] --- maven-compiler-plugin:3.7.0:compile (default-compile) @ dpruntime-tests-api-components --- [INFO] No sources to compile [INFO] [INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ dpruntime-tests-api-components --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 352 resources [INFO] [INFO] --- maven-compiler-plugin:3.7.0:testCompile (default-testCompile) @ dpruntime-tests-api-components --- [INFO] Nothing to compile - all classes are up to date [INFO] [INFO] --- maven-surefire-plugin:2.12.4:test (default-test) @ dpruntime-tests-api-components --- [INFO] Surefire report directory: /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/target/surefire-reports ------------------------------------------------------- T E S T S ------------------------------------------------------- Running CucumberTest @sample-ds Feature: inject all Amazon datasets @sample-ds @component-ds @AmazonKinesis-ds @smoke-ds Scenario: inject AmazonKinesisSTS # features/preparation/datasets/AmazonDatasetPreparation.feature:5 WARNING: An illegal reflective access operation has occurred WARNING: Illegal reflective access by cucumber.deps.com.thoughtworks.xstream.core.util.Fields (file:/root/.m2/repository/io/cucumber/cucumber-jvm-deps/1.0.6/cucumber-jvm-deps-1.0.6.jar) to field java.util.TreeMap.comparator WARNING: Please consider reporting this to the maintainers of cucumber.deps.com.thoughtworks.xstream.core.util.Fields WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations WARNING: All illegal access operations will be denied in a future release SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". SLF4J: Defaulting to no-operation (NOP) logger implementation SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details. When user adds AmazonKinesisSTS dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @smoke-ds @AmazonKinesis-ds Scenario: inject AmazonKinesis # features/preparation/datasets/AmazonDatasetPreparation.feature:10 When user adds AmazonKinesis dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @AmazonDynamoDB-ds @smoke-ds Scenario: inject AmazonDynamoDB # features/preparation/datasets/AmazonDatasetPreparation.feature:14 When user adds AmazonDynamoDB dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @AmazonDynamoDB_AssumeRole-ds @smoke-ds Scenario: inject AmazonDynamoDB_AssumeRole with Table # features/preparation/datasets/AmazonDatasetPreparation.feature:18 When user adds AmazonDynamoDB_AssumeRole dataset with Table configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AmazonS3-ds @smoke-ds Scenario: inject AmazonS3 with Avro # features/preparation/datasets/AmazonDatasetPreparation.feature:22 When user adds AmazonS3 dataset with Avro configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AmazonS3-ds @smoke-ds Scenario: inject AmazonS3 with Excel # features/preparation/datasets/AmazonDatasetPreparation.feature:26 When user adds AmazonS3 dataset with Excel configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AmazonS3-ds @smoke-ds Scenario: inject AmazonS3 with Parquet # features/preparation/datasets/AmazonDatasetPreparation.feature:30 When user adds AmazonS3 dataset with Parquet configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AmazonS3-ds Scenario: inject AmazonS3 with Csv # features/preparation/datasets/AmazonDatasetPreparation.feature:34 When user adds AmazonS3 dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AmazonS3-ds Scenario: inject AmazonS3 with Json # features/preparation/datasets/AmazonDatasetPreparation.feature:38 When user adds AmazonS3 dataset with Json configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds Feature: inject all Azure service datasets @sample-ds @component-ds @AzureEventHubs-ds @smoke-ds Scenario: inject AzureEventHubs with CSV # features/preparation/datasets/AzureDatasetPreparation.feature:9 When user adds AzureEventHubs dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureEventHubs-ds @smoke-ds Scenario: inject AzureEventHubs with AVRO # features/preparation/datasets/AzureDatasetPreparation.feature:13 When user adds AzureEventHubs dataset with Avro configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureEventHubs-ds @smoke-ds Scenario: inject AzureEventHubs with JSON # features/preparation/datasets/AzureDatasetPreparation.feature:17 When user adds AzureEventHubs dataset with Json configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureEventHubs-ds @smoke-ds Scenario: inject AzureEventHubs with TEXT # features/preparation/datasets/AzureDatasetPreparation.feature:21 When user adds AzureEventHubs dataset with Text configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureCosmosDB-ds @smoke-ds Scenario: inject AzureCosmosDB # features/preparation/datasets/AzureDatasetPreparation.feature:25 When user adds AzureCosmosDB dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @AzureDataLakeStorageGen2-ds @smoke-ds Scenario: inject AzureDataLakeStorageGen2 with Avro # features/preparation/datasets/AzureDatasetPreparation.feature:29 When user adds AzureDataLakeStorageGen2 dataset with Avro configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureDataLakeStorageGen2-ds @smoke-ds Scenario: inject AzureDataLakeStorageGen2 with Json # features/preparation/datasets/AzureDatasetPreparation.feature:33 When user adds AzureDataLakeStorageGen2 dataset with Json configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureDataLakeStorageGen2-ds Scenario: inject AzureDataLakeStorageGen2 with Parquet # features/preparation/datasets/AzureDatasetPreparation.feature:37 When user adds AzureDataLakeStorageGen2 dataset with Parquet configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureDataLakeStorageGen2-ds @smoke-ds Scenario: inject AzureDataLakeStorageGen2 with Csv # features/preparation/datasets/AzureDatasetPreparation.feature:41 When user adds AzureDataLakeStorageGen2 dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureDataLakeStorageGen2-ds @smoke-ds Scenario: inject AzureDataLakeStorageGen2 with Tab-Csv # features/preparation/datasets/AzureDatasetPreparation.feature:45 When user adds AzureDataLakeStorageGen2 dataset with Tab-Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureDataLakeStorageGen2-ds @smoke-ds Scenario: inject AzureDataLakeStorageGen2 with OtherSettings-Csv # features/preparation/datasets/AzureDatasetPreparation.feature:49 When user adds AzureDataLakeStorageGen2 dataset with OtherSettings-Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureDataLakeStorageGen2SAS-ds @smoke-ds Scenario: inject AzureDataLakeStorageGen2SAS with Avro # features/preparation/datasets/AzureDatasetPreparation.feature:53 When user adds AzureDataLakeStorageGen2SAS dataset with Avro configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureDataLakeStorageGen2SAS-ds @smoke-ds Scenario: inject AzureDataLakeStorageGen2SAS with Json # features/preparation/datasets/AzureDatasetPreparation.feature:57 When user adds AzureDataLakeStorageGen2SAS dataset with Json configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureDataLakeStorageGen2SAS-ds @smoke-ds Scenario: inject AzureDataLakeStorageGen2SAS with Parquet # features/preparation/datasets/AzureDatasetPreparation.feature:61 When user adds AzureDataLakeStorageGen2SAS dataset with Parquet configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureDataLakeStorageGen2SAS-ds @smoke-ds Scenario: inject AzureDataLakeStorageGen2SAS with Csv # features/preparation/datasets/AzureDatasetPreparation.feature:65 When user adds AzureDataLakeStorageGen2SAS dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureBlobStorage-ds @smoke-ds Scenario: inject AzureBlobStorage with Avro # features/preparation/datasets/AzureDatasetPreparation.feature:89 When user adds AzureBlobStorage dataset with Avro configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureBlobStorage-ds Scenario: inject AzureBlobStorage with Parquet # features/preparation/datasets/AzureDatasetPreparation.feature:93 When user adds AzureBlobStorage dataset with Parquet configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureBlobStorage-ds @smoke-ds Scenario: inject AzureBlobStorage with Csv # features/preparation/datasets/AzureDatasetPreparation.feature:97 When user adds AzureBlobStorage dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureBlobStorage-ds Scenario: inject AzureBlobStorage with Excel # features/preparation/datasets/AzureDatasetPreparation.feature:101 When user adds AzureBlobStorage dataset with Excel configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureBlobStorage-ds Scenario: inject AzureBlobStorage with Json # features/preparation/datasets/AzureDatasetPreparation.feature:105 When user adds AzureBlobStorage dataset with Json configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureBlobStorageSAS-ds @smoke-ds Scenario: inject AzureBlobStorageSAS with Avro # features/preparation/datasets/AzureDatasetPreparation.feature:109 When user adds AzureBlobStorageSAS dataset with Avro configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureBlobStorageSAS-ds @smoke-ds Scenario: inject AzureBlobStorageSAS with Parquet # features/preparation/datasets/AzureDatasetPreparation.feature:113 When user adds AzureBlobStorageSAS dataset with Parquet configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureBlobStorageSAS-ds @smoke-ds Scenario: inject AzureBlobStorageSAS with Csv # features/preparation/datasets/AzureDatasetPreparation.feature:117 When user adds AzureBlobStorageSAS dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureBlobStorageSAS-ds @smoke-ds Scenario: inject AzureBlobStorageSAS with Excel # features/preparation/datasets/AzureDatasetPreparation.feature:121 When user adds AzureBlobStorageSAS dataset with Excel configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds Feature: inject all common datasets @sample-ds @component-ds @Test-ds Scenario: inject Test dataset # features/preparation/datasets/CommonDatasetPreparation.feature:5 When user adds Test dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @AzureDyanmic165-ds @smoke-ds Scenario: inject TestAzureDynamic365 dataset # features/preparation/datasets/CommonDatasetPreparation.feature:9 When user adds TestAzureDynamic365 dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @AzureDataLakeStorageGen2-ds @smoke-ds Scenario: inject TestEscape dataset # features/preparation/datasets/CommonDatasetPreparation.feature:13 When user adds TestEscape dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @GooglePubSub-ds @smoke-ds Scenario: inject TestGooglePubSub dataset # features/preparation/datasets/CommonDatasetPreparation.feature:17 When user adds TestGooglePubSub dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @GooglePubSub-ds @smoke-ds Scenario: inject TestGooglePubSub dataset # features/preparation/datasets/CommonDatasetPreparation.feature:21 When user adds TestGooglePubSub dataset with CSV configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds @smoke-ds Scenario: inject TestCommonIO dataset # features/preparation/datasets/CommonDatasetPreparation.feature:25 When user adds TestCommonIO dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @NetSuite-ds @smoke-ds Scenario: inject TestNetsuite dataset # features/preparation/datasets/CommonDatasetPreparation.feature:29 When user adds TestNetsuite dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @Salesforce-ds @smoke-ds Scenario: inject TestSalesforce dataset # features/preparation/datasets/CommonDatasetPreparation.feature:33 When user adds TestSalesforce dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @AzureEventHubs-ds @smoke-ds Scenario: inject TestEventHub Avro dataset # features/preparation/datasets/CommonDatasetPreparation.feature:37 When user adds TestEventHub dataset with Avro configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @AzureEventHubs-ds @smoke-ds Scenario: inject TestEventHub TEXT dataset # features/preparation/datasets/CommonDatasetPreparation.feature:41 When user adds TestEventHub dataset with Text configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @ElasticSearch-ds @smoke-ds Scenario: inject ElasticSearch dataset # features/preparation/datasets/CommonDatasetPreparation.feature:45 When user adds ElasticSearch dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @Kafka-ds @smoke-ds Scenario: inject Kafka dataset with Csv # features/preparation/datasets/CommonDatasetPreparation.feature:49 When user adds Kafka dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Kafka-ds @smoke-ds Scenario: inject Kafka dataset with Json # features/preparation/datasets/CommonDatasetPreparation.feature:53 When user adds Kafka dataset with Json configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Kafka-ds Scenario: inject Kafka dataset with Avro # features/preparation/datasets/CommonDatasetPreparation.feature:57 When user adds Kafka dataset with Avro configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @REST-ds @smoke-ds @SourceIsTarget-ds Scenario: inject REST dataset with Batch # features/preparation/datasets/CommonDatasetPreparation.feature:61 When user adds REST dataset with Batch configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @REST-ds @smoke-ds @SourceIsTarget-ds Scenario: inject REST dataset with Streaming # features/preparation/datasets/CommonDatasetPreparation.feature:65 When user adds REST dataset with Streaming configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @REST-ds @smoke-ds @SourceIsTarget-ds Scenario: inject REST dataset with Header-Parameter # features/preparation/datasets/CommonDatasetPreparation.feature:69 When user adds REST dataset with Header-Parameter configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @REST-ds @smoke-ds @SourceIsTarget-ds Scenario: inject REST dataset with URL-Parameter # features/preparation/datasets/CommonDatasetPreparation.feature:73 When user adds REST dataset with URL-Parameter configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @REST-ds @smoke-ds Scenario: inject REST_HTTPBin dataset with Path-Parameter # features/preparation/datasets/CommonDatasetPreparation.feature:77 When user adds REST_HTTPBin dataset with Path-Parameter configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @REST-ds @smoke-ds Scenario: inject REST_HTTPBin dataset with Body-Parameter # features/preparation/datasets/CommonDatasetPreparation.feature:81 When user adds REST_HTTPBin dataset with Body-Parameter configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @REST-ds @smoke-ds Scenario: inject REST_Authentication_Basic dataset with Basic_Auth # features/preparation/datasets/CommonDatasetPreparation.feature:85 When user adds REST_Authentication_Basic dataset with Basic_Auth configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @REST-ds @smoke-ds Scenario: inject REST_Authentication_Digest dataset with Digest_Auth # features/preparation/datasets/CommonDatasetPreparation.feature:89 When user adds REST_Authentication_Digest dataset with Digest_Auth configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @MongoDB-ds @smoke-ds Scenario: inject MongoDB dataset with Collection-Json # features/preparation/datasets/CommonDatasetPreparation.feature:93 When user adds MongoDB dataset with Collection-Json configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Pulsar-ds @smoke-ds Scenario: inject Pulsar dataset with Csv # features/preparation/datasets/CommonDatasetPreparation.feature:97 When user adds Pulsar dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Marketo-ds @smoke-ds Scenario: inject Marketo dataset with People # features/preparation/datasets/CommonDatasetPreparation.feature:101 When user adds Marketo dataset with People configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Salesforce-ds @smoke-ds Scenario: inject Salesforce dataset with Module # features/preparation/datasets/CommonDatasetPreparation.feature:105 When user adds Salesforce dataset with Module configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Salesforce-ds Scenario: inject Salesforce dataset with Query # features/preparation/datasets/CommonDatasetPreparation.feature:109 When user adds Salesforce dataset with Query configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Workday-ds @smoke-ds Scenario: inject Workday dataset with Raas # features/preparation/datasets/CommonDatasetPreparation.feature:115 When user adds Workday dataset with RaaS configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Workday-ds @smoke-ds Scenario: inject Workday dataset with WSSA # features/preparation/datasets/CommonDatasetPreparation.feature:119 When user adds Workday dataset with WSSA configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @NetSuite-ds @smoke-ds Scenario: inject NetSuite # features/preparation/datasets/CommonDatasetPreparation.feature:123 When user adds NetSuite dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @HDFS-ds @smoke-ds Scenario: inject HDFS dataset with Csv # features/preparation/datasets/CommonDatasetPreparation.feature:127 When user adds HDFS dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @FTP-ds @smoke-ds Scenario: inject FTP dataset with Csv # features/preparation/datasets/CommonDatasetPreparation.feature:131 When user adds FTP dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @SFTP-ds @smoke-ds Scenario: inject SFTP dataset with Csv # features/preparation/datasets/CommonDatasetPreparation.feature:135 When user adds SFTP dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @LocalConnection-ds @smoke-ds Scenario: inject LocalConnection dataset with Csv # features/preparation/datasets/CommonDatasetPreparation.feature:139 When user adds LocalConnection dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Zendesk-ds @smoke-ds Scenario: inject Zendesk dataset with Tickets # features/preparation/datasets/CommonDatasetPreparation.feature:143 When user adds Zendesk dataset with Tickets-Create configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Zendesk-ds @smoke-ds Scenario: inject Zendesk dataset with Tickets # features/preparation/datasets/CommonDatasetPreparation.feature:147 When user adds Zendesk dataset with Requests-Create configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Zendesk-ds @smoke-ds Scenario: inject TestZendesk dataset # features/preparation/datasets/CommonDatasetPreparation.feature:151 When user adds TestZendesk dataset with Requests configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Zendesk-ds @smoke-ds Scenario: inject TestZendesk dataset # features/preparation/datasets/CommonDatasetPreparation.feature:155 When user adds TestZendesk dataset with Tickets configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Couchbase-ds @smoke-ds Scenario: inject Couchbase dataset with JSON # features/preparation/datasets/CommonDatasetPreparation.feature:159 When user adds Couchbase dataset with JSON configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Kudu-ds @smoke-ds Scenario: inject Kudu dataset # features/preparation/datasets/CommonDatasetPreparation.feature:163 When user adds Kudu dataset without additional configuration # AddDatasetStep.userAddsDataset(String) @sample-ds @component-ds @Box-ds @smoke-ds @SourceIsTarget Scenario: inject BoxUId dataset with Csv # features/preparation/datasets/CommonDatasetPreparation.feature:167 Box private Key length:1834 Box private Key Passphrase length:32 When user adds Box_User_Id dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Box-ds @smoke-ds Scenario: inject BoxUId dataset with Avro # features/preparation/datasets/CommonDatasetPreparation.feature:171 Box private Key length:1834 Box private Key Passphrase length:32 When user adds Box_User_Id dataset with Avro configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Box-ds @smoke-ds Scenario: inject BoxUId dataset with Json # features/preparation/datasets/CommonDatasetPreparation.feature:175 Box private Key length:1834 Box private Key Passphrase length:32 When user adds Box_User_Id dataset with Json configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Box-ds @smoke-ds Scenario: inject BoxUId dataset with Excel # features/preparation/datasets/CommonDatasetPreparation.feature:179 Box private Key length:1834 Box private Key Passphrase length:32 When user adds Box_User_Id dataset with Excel configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Box-ds @smoke-ds Scenario: inject BoxUId dataset with Excel-HTML-Read # features/preparation/datasets/CommonDatasetPreparation.feature:183 Box private Key length:1834 Box private Key Passphrase length:32 When user adds Box_User_Id dataset with Excel-HTML-Read configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Box-ds @smoke-ds @SourceIsTarget Scenario: inject BoxEId dataset with Csv # features/preparation/datasets/CommonDatasetPreparation.feature:187 Box private Key length:1834 Box private Key Passphrase length:32 When user adds Box_Enterprise_Id dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Box-ds @smoke-ds Scenario: inject BoxEId dataset with Avro # features/preparation/datasets/CommonDatasetPreparation.feature:191 Box private Key length:1834 Box private Key Passphrase length:32 When user adds Box_Enterprise_Id dataset with Avro configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Box-ds @smoke-ds Scenario: inject BoxEId dataset with Json # features/preparation/datasets/CommonDatasetPreparation.feature:195 Box private Key length:1834 Box private Key Passphrase length:32 When user adds Box_Enterprise_Id dataset with Json configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Box-ds @smoke-ds Scenario: inject BoxEId dataset with Excel # features/preparation/datasets/CommonDatasetPreparation.feature:199 Box private Key length:1834 Box private Key Passphrase length:32 When user adds Box_Enterprise_Id dataset with Excel configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @Box-ds @smoke-ds Scenario: inject BoxEId dataset with Excel-HTML-Read # features/preparation/datasets/CommonDatasetPreparation.feature:203 Box private Key length:1834 Box private Key Passphrase length:32 When user adds Box_Enterprise_Id dataset with Excel-HTML-Read configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @Database-ds Feature: inject all Database datasets @sample-ds @Database-ds @component-ds @PostgreSQL-ds @smoke-ds Scenario: inject PostgreSQL with Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:5 When user adds database dataset PostgreSQL with Table configuration # AddDatasetStep.userAddsDatabaseDataset(String,String) @sample-ds @Database-ds @component-ds @PostgreSQL-ds @smoke-ds Scenario: inject PostgreSQL with the Basic of authentication type and Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:9 When user adds database authenticationType Basic dataset PostgreSQL with Table configuration # AddDatasetStep.userAddsDatabaseWithAuthType(String,String,String) @sample-ds @Database-ds @component-ds @Aurora-ds @smoke-ds Scenario: inject Aurora with Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:13 When user adds database dataset Aurora with Table configuration # AddDatasetStep.userAddsDatabaseDataset(String,String) @sample-ds @Database-ds @component-ds @MariaDB-ds @smoke-ds Scenario: inject MariaDB with Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:17 When user adds database dataset MariaDB with Table configuration # AddDatasetStep.userAddsDatabaseDataset(String,String) @sample-ds @Database-ds @component-ds @MariaDB-ds @smoke-ds Scenario: inject MariaDB with the Basic of authentication type and Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:21 When user adds database authenticationType Basic dataset MariaDB with Table configuration # AddDatasetStep.userAddsDatabaseWithAuthType(String,String,String) @sample-ds @Database-ds @component-ds @MSSQL-ds @smoke-ds Scenario: inject MSSQL with Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:25 When user adds database dataset MSSQL with Table configuration # AddDatasetStep.userAddsDatabaseDataset(String,String) @sample-ds @Database-ds @component-ds @MSSQL-ds @smoke-ds Scenario: inject MSSQL with the Basic of authentication type and Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:29 When user adds database authenticationType Basic dataset MSSQL with Table configuration # AddDatasetStep.userAddsDatabaseWithAuthType(String,String,String) @sample-ds @Database-ds @component-ds @MySql-ds @smoke-ds Scenario: inject MySQL with Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:33 When user adds database dataset MySQL with Table configuration # AddDatasetStep.userAddsDatabaseDataset(String,String) @sample-ds @Database-ds @component-ds @MySql-ds @smoke-ds Scenario: inject MySQL with the Basic of authentication type and Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:37 When user adds database authenticationType Basic dataset MySQL with Table configuration # AddDatasetStep.userAddsDatabaseWithAuthType(String,String,String) @sample-ds @Database-ds @component-ds @RedShift-ds @smoke-ds Scenario: inject Redshift with Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:41 When user adds database dataset Redshift with Table configuration # AddDatasetStep.userAddsDatabaseDataset(String,String) @sample-ds @Database-ds @component-ds @Snowflake-ds @smoke-ds Scenario: inject Snowflake with Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:45 When user adds database dataset Snowflake with Table configuration # AddDatasetStep.userAddsDatabaseDataset(String,String) @sample-ds @Database-ds @component-ds @Snowflake-ds @smoke-ds Scenario: inject Snowflake with the Basic of authentication type and Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:49 When user adds database authenticationType Basic dataset Snowflake with Table configuration # AddDatasetStep.userAddsDatabaseWithAuthType(String,String,String) @sample-ds @Database-ds @component-ds @Snowflake-ds @smoke-ds @component-ds @AzureSynapse-ds @smoke-ds Scenario: inject AzureSynapse with Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:57 When user adds database dataset AzureSynapse with Table configuration # AddDatasetStep.userAddsDatabaseDataset(String,String) @sample-ds @Database-ds @component-ds @Derby-ds @smoke-ds Scenario: inject Derby with Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:61 When user adds database dataset Derby with Table configuration # AddDatasetStep.userAddsDatabaseDataset(String,String) @sample-ds @Database-ds @component-ds @SingleStore-ds @smoke-ds Scenario: inject SingleStore with Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:65 When user adds database dataset SingleStore with Table configuration # AddDatasetStep.userAddsDatabaseDataset(String,String) @sample-ds @Database-ds @component-ds @DeltaLake-ds @smoke-ds Scenario: inject DeltaLake with Table # features/preparation/datasets/DatabaseDatasetPreparation.feature:69 When user adds database dataset DeltaLake with Table configuration # AddDatasetStep.userAddsDatabaseDataset(String,String) @sample-ds Feature: inject all Google service datasets @sample-ds @component-ds @GoogleBigQuery-ds @smoke-ds Scenario: inject GoogleBigQuery dataset with Table # features/preparation/datasets/GoogleDatasetPreparation.feature:5 When user adds GoogleBigQuery dataset with Table configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GooglePubSub-ds @smoke-ds Scenario: inject GooglePubSub dataset with Csv # features/preparation/datasets/GoogleDatasetPreparation.feature:9 When user adds GooglePubSub dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GooglePubSub-ds Scenario: inject GooglePubSub dataset with Text # features/preparation/datasets/GoogleDatasetPreparation.feature:13 When user adds GooglePubSub dataset with Text configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GooglePubSub-ds @smoke-ds Scenario: inject GooglePubSub dataset with Avro # features/preparation/datasets/GoogleDatasetPreparation.feature:17 When user adds GooglePubSub dataset with Avro configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GooglePubSub-ds Scenario: inject GooglePubSub dataset with JSON # features/preparation/datasets/GoogleDatasetPreparation.feature:21 When user adds GooglePubSub dataset with JSON configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds @smoke-ds Scenario: inject GoogleCloudStorage dataset with Csv # features/preparation/datasets/GoogleDatasetPreparation.feature:25 When user adds GoogleCloudStorage dataset with Csv configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with Path # features/preparation/datasets/GoogleDatasetPreparation.feature:29 When user adds GoogleCloudStorage dataset with Path configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with CommonIO-JSON # features/preparation/datasets/GoogleDatasetPreparation.feature:33 When user adds GoogleCloudStorage dataset with CommonIO-JSON configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with CommonIO-AVRO # features/preparation/datasets/GoogleDatasetPreparation.feature:37 When user adds GoogleCloudStorage dataset with CommonIO-AVRO configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with CommonIO-Excel97 # features/preparation/datasets/GoogleDatasetPreparation.feature:41 When user adds GoogleCloudStorage dataset with CommonIO-Excel97 configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with CommonIO-Excel2007 # features/preparation/datasets/GoogleDatasetPreparation.feature:45 When user adds GoogleCloudStorage dataset with CommonIO-Excel2007 configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with CommonIO-CSV-LinuxType # features/preparation/datasets/GoogleDatasetPreparation.feature:49 When user adds GoogleCloudStorage dataset with CommonIO-CSV-LinuxType configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with CommonIO-CSV-WindowsType # features/preparation/datasets/GoogleDatasetPreparation.feature:53 When user adds GoogleCloudStorage dataset with CommonIO-CSV-WindowsType configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with CommonIO-CSV-Header # features/preparation/datasets/GoogleDatasetPreparation.feature:57 When user adds GoogleCloudStorage dataset with CommonIO-CSV-Header configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with CommonIO-CSV-ISO-8859-15 # features/preparation/datasets/GoogleDatasetPreparation.feature:61 When user adds GoogleCloudStorage dataset with CommonIO-CSV-ISO-8859-15 configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with CommonIO-CSV-PipeFieldSeparator # features/preparation/datasets/GoogleDatasetPreparation.feature:65 When user adds GoogleCloudStorage dataset with CommonIO-CSV-PipeFieldSeparator configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with CommonIO-CSV-SemicolonFieldSeparator # features/preparation/datasets/GoogleDatasetPreparation.feature:69 When user adds GoogleCloudStorage dataset with CommonIO-CSV-SemicolonFieldSeparator configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with CommonIO-CSV-Advanced-EscapeCharacter # features/preparation/datasets/GoogleDatasetPreparation.feature:73 When user adds GoogleCloudStorage dataset with CommonIO-CSV-Advanced-EscapeCharacter configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds @component-ds @GoogleCloudStorage-ds Scenario: inject GoogleCloudStorage dataset with CommonIO-CSV-Advanced-QuoteCharacter # features/preparation/datasets/GoogleDatasetPreparation.feature:77 When user adds GoogleCloudStorage dataset with CommonIO-CSV-Advanced-QuoteCharacter configuration # AddDatasetStep.userAddsDataset(String,String) @sample-ds Feature: save datasets' id @sample-ds @saveDsId Scenario: save datasets' id # features/preparation/datasets/WeSaveDatasetId.feature:5 When add a json file to contain all datasets' id # UtilStep.saveDatasetIds() 119 Scenarios (119 passed) 119 Steps (119 passed) 4m15.438s Tests run: 119, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 256.297 sec Results : Tests run: 119, Failures: 0, Errors: 0, Skipped: 0 [INFO] ------------------------------------------------------------------------ [INFO] BUILD SUCCESS [INFO] ------------------------------------------------------------------------ [INFO] Total time: 04:24 min [INFO] Finished at: 2022-11-14T13:55:34Z [INFO] ------------------------------------------------------------------------ [Pipeline] } [Pipeline] // script [Pipeline] } [Pipeline] // stage [Pipeline] stage [Pipeline] { (TEST) [Pipeline] script [Pipeline] { [Pipeline] sh [WARNING] [WARNING] Some problems were encountered while building the effective settings [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] 'servers.server.id' must be unique but found duplicate server with id $id @ /root/.m2/settings.xml [WARNING] [INFO] Scanning for projects... [INFO] [INFO] ---------< org.talend.qa.test:dpruntime-tests-api-components >---------- [INFO] Building dpruntime-tests-api-components 1.0.0-SNAPSHOT [INFO] --------------------------------[ jar ]--------------------------------- [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ dpruntime-tests-api-components --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] skip non existing resourceDirectory /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/src/main/resources [INFO] [INFO] --- maven-compiler-plugin:3.7.0:compile (default-compile) @ dpruntime-tests-api-components --- [INFO] No sources to compile [INFO] [INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ dpruntime-tests-api-components --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 352 resources [INFO] [INFO] --- maven-compiler-plugin:3.7.0:testCompile (default-testCompile) @ dpruntime-tests-api-components --- [INFO] Nothing to compile - all classes are up to date [INFO] [INFO] --- maven-surefire-plugin:2.12.4:test (default-test) @ dpruntime-tests-api-components --- [INFO] Surefire report directory: /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/target/surefire-reports ------------------------------------------------------- T E S T S ------------------------------------------------------- Running CucumberTest @component Feature: test AmazonDynamoDB components @component @smoke @AmazonDynamoDB Scenario: test -> DynamoDB # features/connectors/Amazon/AmazonDynamoDB.feature:5 WARNING: An illegal reflective access operation has occurred WARNING: Illegal reflective access by cucumber.deps.com.thoughtworks.xstream.core.util.Fields (file:/root/.m2/repository/io/cucumber/cucumber-jvm-deps/1.0.6/cucumber-jvm-deps-1.0.6.jar) to field java.util.TreeMap.comparator WARNING: Please consider reporting this to the maintainers of cucumber.deps.com.thoughtworks.xstream.core.util.Fields WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations WARNING: All illegal access operations will be denied in a future release Nov 14, 2022 1:55:45 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 1:55:46 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 1:55:46 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 99679d7b-a05a-4543-8fcf-4518a6ac5160 for {dataset_id} SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". SLF4J: Defaulting to no-operation (NOP) logger implementation SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details. Nov 14, 2022 1:55:52 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonDynamoDB is : 201 When user adds pipeline using AmazonDynamoDB without dataset config for output component with Insert config # AddPipelineStep.addPipelineWithSetting(String,String,String) Nov 14, 2022 1:55:52 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AmazonDynamoDB Scenario: DynamoDB -> test # features/connectors/Amazon/AmazonDynamoDB.feature:11 Nov 14, 2022 1:56:13 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 1:56:13 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 99679d7b-a05a-4543-8fcf-4518a6ac5160 for {dataset_id} Nov 14, 2022 1:56:14 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonDynamoDB is : 201 When user adds pipeline using AmazonDynamoDB without dataset config for input component with Scan config # AddPipelineStep.addPipelineWithSetting(String,String,String) Nov 14, 2022 1:56:14 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AmazonDynamoDB_AssumeRole Scenario: test -> DynamoDB_AssumeRole # features/connectors/Amazon/AmazonDynamoDB.feature:23 Nov 14, 2022 1:56:29 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 1:56:29 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 1:56:29 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: c4dbbacc-11d5-4584-ae08-a46449296b09 for {dataset_id} Nov 14, 2022 1:56:29 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonDynamoDB_AssumeRole is : 201 When user adds pipeline using AmazonDynamoDB_AssumeRole with dataset Table config for output component with Insert config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 1:56:29 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AmazonDynamoDB_AssumeRole Scenario: DynamoDB_AssumeRole Query -> test # features/connectors/Amazon/AmazonDynamoDB.feature:29 Nov 14, 2022 1:56:45 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 1:56:45 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: c4dbbacc-11d5-4584-ae08-a46449296b09 for {dataset_id} Nov 14, 2022 1:56:45 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonDynamoDB_AssumeRole is : 201 When user adds pipeline using AmazonDynamoDB_AssumeRole with dataset Table config for input component with Query config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 1:56:45 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test AmazonKinesis components @component @smoke @AmazonKinesis Scenario: Kinesis with STS -> test # features/connectors/Amazon/AmazonKinesis.feature:5 Nov 14, 2022 1:57:00 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 1:57:00 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 285604cd-debe-4bee-9d5b-c992d907eca0 for {dataset_id} Nov 14, 2022 1:57:00 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonKinesisSTS is : 201 When user adds pipeline using AmazonKinesisSTS without dataset config for input component with Latest config # AddPipelineStep.addPipelineWithSetting(String,String,String) Nov 14, 2022 1:57:01 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 1:58:34 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component Feature: test AmazonS3 components @component @smoke @AmazonS3 Scenario: test dataset -> S3 with Excel # features/connectors/Amazon/AmazonS3.feature:5 Nov 14, 2022 1:58:37 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 1:58:37 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 1:58:37 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 7664d690-0b5d-459e-b023-761c0eb63ccc for {dataset_id} Nov 14, 2022 1:58:37 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonS3 is : 201 When user adds pipeline using AmazonS3 with dataset Excel config for output component with OverwriteMerge config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 1:58:37 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AmazonS3 Scenario: test dataset -> S3 with AVRO # features/connectors/Amazon/AmazonS3.feature:11 Nov 14, 2022 1:59:35 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 1:59:35 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 1:59:35 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: e3f879b5-530f-4a7e-9412-b36391990120 for {dataset_id} Nov 14, 2022 1:59:36 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonS3 is : 201 When user adds pipeline using AmazonS3 with dataset Avro config for output component with OverwriteMerge config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 1:59:36 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AmazonS3 Scenario: test dataset -> S3 with Parquet # features/connectors/Amazon/AmazonS3.feature:17 Nov 14, 2022 1:59:57 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 1:59:57 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 1:59:57 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 998778e4-6868-427e-bc6f-eeda9c762305 for {dataset_id} Nov 14, 2022 1:59:57 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonS3 is : 201 When user adds pipeline using AmazonS3 with dataset Parquet config for output component with OverwriteMerge config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 1:59:58 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AmazonS3 Scenario: test dataset -> S3 with Csv # features/connectors/Amazon/AmazonS3.feature:23 Nov 14, 2022 2:00:19 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:00:19 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:00:19 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 5adeda06-8d68-4bfe-b097-0053b35b8b2d for {dataset_id} Nov 14, 2022 2:00:19 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonS3 is : 201 When user adds pipeline using AmazonS3 with dataset Csv config for output component with OverwriteMerge config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:00:19 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AmazonS3 Scenario: test dataset -> S3 with Json # features/connectors/Amazon/AmazonS3.feature:29 Nov 14, 2022 2:00:41 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:00:41 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:00:41 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a0aac6a0-4c2e-435c-8cbd-4c1c9dd4a055 for {dataset_id} Nov 14, 2022 2:00:41 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonS3 is : 201 When user adds pipeline using AmazonS3 with dataset Json config for output component with OverwriteMerge config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:00:41 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AmazonS3 Scenario: S3 with Excel -> test # features/connectors/Amazon/AmazonS3.feature:35 Nov 14, 2022 2:01:03 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:01:03 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 7664d690-0b5d-459e-b023-761c0eb63ccc for {dataset_id} Nov 14, 2022 2:01:04 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonS3 is : 201 When user adds pipeline using AmazonS3 with dataset Excel config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:01:04 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AmazonS3 Scenario: S3 with Parquet -> test # features/connectors/Amazon/AmazonS3.feature:41 Nov 14, 2022 2:01:56 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:01:56 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 998778e4-6868-427e-bc6f-eeda9c762305 for {dataset_id} Nov 14, 2022 2:01:56 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonS3 is : 201 When user adds pipeline using AmazonS3 with dataset Parquet config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:01:56 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AmazonS3 Scenario: S3 with Avro -> test # features/connectors/Amazon/AmazonS3.feature:47 Nov 14, 2022 2:02:18 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:02:18 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: e3f879b5-530f-4a7e-9412-b36391990120 for {dataset_id} Nov 14, 2022 2:02:18 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonS3 is : 201 When user adds pipeline using AmazonS3 with dataset Avro config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:02:18 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AmazonS3 Scenario: S3 with Csv -> test # features/connectors/Amazon/AmazonS3.feature:53 Nov 14, 2022 2:02:40 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:02:40 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 5adeda06-8d68-4bfe-b097-0053b35b8b2d for {dataset_id} Nov 14, 2022 2:02:40 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonS3 is : 201 When user adds pipeline using AmazonS3 with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:02:40 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AmazonS3 Scenario: S3 with Json -> test # features/connectors/Amazon/AmazonS3.feature:59 Nov 14, 2022 2:03:01 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:03:01 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a0aac6a0-4c2e-435c-8cbd-4c1c9dd4a055 for {dataset_id} Nov 14, 2022 2:03:01 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AmazonS3 is : 201 When user adds pipeline using AmazonS3 with dataset Json config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:03:01 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test AzureBlobStorage components @component @smoke @AzureBlobStorage Scenario: test data -> Azure Blob Storage with Csv # features/connectors/Azure/AzureBlobStorage.feature:5 Nov 14, 2022 2:03:23 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:03:23 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:03:23 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: f0653044-cd22-419b-93a9-cfc7af0b82ac for {dataset_id} Nov 14, 2022 2:03:23 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorage is : 201 When user adds pipeline using AzureBlobStorage with dataset Csv config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:03:23 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorage Scenario: test data -> Azure Blob Storage with Avro # features/connectors/Azure/AzureBlobStorage.feature:11 Nov 14, 2022 2:03:38 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:03:38 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:03:38 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 02ebbb29-b0b7-4b7e-b0ce-cbefad9485e0 for {dataset_id} Nov 14, 2022 2:03:39 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorage is : 201 When user adds pipeline using AzureBlobStorage with dataset Avro config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:03:39 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorage Scenario: test data -> Azure Blob Storage with Parquet # features/connectors/Azure/AzureBlobStorage.feature:17 Nov 14, 2022 2:03:54 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:03:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:03:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: b5c9162e-9bf1-4c31-9084-10956576a866 for {dataset_id} Nov 14, 2022 2:03:54 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorage is : 201 When user adds pipeline using AzureBlobStorage with dataset Parquet config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:03:54 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorage Scenario: Test -> Azure Blob Storage with Excel # features/connectors/Azure/AzureBlobStorage.feature:23 Nov 14, 2022 2:04:16 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:04:16 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:04:16 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 3e801090-3bc4-4125-99ad-44c0d87f8056 for {dataset_id} Nov 14, 2022 2:04:16 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorage is : 201 When user adds pipeline using AzureBlobStorage with dataset Excel config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:04:16 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorage Scenario: Test -> Azure Blob Storage with Json # features/connectors/Azure/AzureBlobStorage.feature:28 Nov 14, 2022 2:04:37 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:04:37 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:04:37 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a7755234-25f5-48db-bd19-78d833279402 for {dataset_id} Nov 14, 2022 2:04:38 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorage is : 201 When user adds pipeline using AzureBlobStorage with dataset Json config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:04:38 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorage Scenario: Azure Blob Storage with Csv -> test # features/connectors/Azure/AzureBlobStorage.feature:34 Nov 14, 2022 2:04:53 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:04:53 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: f0653044-cd22-419b-93a9-cfc7af0b82ac for {dataset_id} Nov 14, 2022 2:04:53 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorage is : 201 When user adds pipeline using AzureBlobStorage with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:04:53 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureBlobStorage Scenario: Azure Blob Storage with Avro -> test # features/connectors/Azure/AzureBlobStorage.feature:40 Nov 14, 2022 2:05:15 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:05:15 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 02ebbb29-b0b7-4b7e-b0ce-cbefad9485e0 for {dataset_id} Nov 14, 2022 2:05:15 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorage is : 201 When user adds pipeline using AzureBlobStorage with dataset Avro config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:05:15 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorage Scenario: Azure Blob Storage with Parquet -> test # features/connectors/Azure/AzureBlobStorage.feature:46 Nov 14, 2022 2:05:36 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:05:36 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: b5c9162e-9bf1-4c31-9084-10956576a866 for {dataset_id} Nov 14, 2022 2:05:37 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorage is : 201 When user adds pipeline using AzureBlobStorage with dataset Parquet config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:05:37 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorage Scenario: Azure Blob Storage with Excel -> test # features/connectors/Azure/AzureBlobStorage.feature:52 Nov 14, 2022 2:05:58 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:05:58 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 3e801090-3bc4-4125-99ad-44c0d87f8056 for {dataset_id} Nov 14, 2022 2:05:58 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorage is : 201 When user adds pipeline using AzureBlobStorage with dataset Excel config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:05:58 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorage Scenario: Azure Blob Storage with Json -> test # features/connectors/Azure/AzureBlobStorage.feature:58 Nov 14, 2022 2:06:20 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:06:20 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a7755234-25f5-48db-bd19-78d833279402 for {dataset_id} Nov 14, 2022 2:06:20 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorage is : 201 When user adds pipeline using AzureBlobStorage with dataset Json config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:06:20 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test AzureBlobStorageSAS components @component @AzureBlobStorageSAS Scenario: test data -> Azure Blob Storage of SAS with Csv # features/connectors/Azure/AzureBlobStorageSAS.feature:5 Nov 14, 2022 2:06:42 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:06:42 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:06:42 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 7e7d0b57-15e3-4a71-98d6-a63f5a9ca635 for {dataset_id} Nov 14, 2022 2:06:42 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorageSAS is : 201 When user adds pipeline using AzureBlobStorageSAS with dataset Csv config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:06:42 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorageSAS Scenario: test data -> Azure Blob Storage of SAS with Avro # features/connectors/Azure/AzureBlobStorageSAS.feature:11 Nov 14, 2022 2:06:57 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:06:57 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:06:57 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 38edc1c3-b4a0-46d1-bef7-b6fb340d35c7 for {dataset_id} Nov 14, 2022 2:06:57 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorageSAS is : 201 When user adds pipeline using AzureBlobStorageSAS with dataset Avro config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:06:58 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureBlobStorageSAS Scenario: test data -> Azure Blob Storage of SAS with Parquet # features/connectors/Azure/AzureBlobStorageSAS.feature:17 Nov 14, 2022 2:07:13 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:07:13 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:07:13 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 7a753ee2-2b38-4332-8a76-6f25b581b26d for {dataset_id} Nov 14, 2022 2:07:13 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorageSAS is : 201 When user adds pipeline using AzureBlobStorageSAS with dataset Parquet config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:07:13 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorageSAS Scenario: Test -> Azure Blob Storage of SAS with Excel # features/connectors/Azure/AzureBlobStorageSAS.feature:22 Nov 14, 2022 2:07:35 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:07:35 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:07:35 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 3b7280dd-cf5d-44d2-a30c-9a74ccc4a8d9 for {dataset_id} Nov 14, 2022 2:07:35 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorageSAS is : 201 When user adds pipeline using AzureBlobStorageSAS with dataset Excel config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:07:35 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorageSAS Scenario: Azure Blob Storage of SAS with Csv -> test # features/connectors/Azure/AzureBlobStorageSAS.feature:28 Nov 14, 2022 2:07:50 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:07:50 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 7e7d0b57-15e3-4a71-98d6-a63f5a9ca635 for {dataset_id} Nov 14, 2022 2:07:50 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorageSAS is : 201 When user adds pipeline using AzureBlobStorageSAS with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:07:50 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureBlobStorageSAS Scenario: Azure Blob Storage of SAS with Avro -> test # features/connectors/Azure/AzureBlobStorageSAS.feature:34 Nov 14, 2022 2:08:12 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:08:12 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 38edc1c3-b4a0-46d1-bef7-b6fb340d35c7 for {dataset_id} Nov 14, 2022 2:08:12 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorageSAS is : 201 When user adds pipeline using AzureBlobStorageSAS with dataset Avro config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:08:12 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorageSAS Scenario: Azure Blob Storage of SAS with Parquet -> test # features/connectors/Azure/AzureBlobStorageSAS.feature:40 Nov 14, 2022 2:08:34 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:08:34 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 7a753ee2-2b38-4332-8a76-6f25b581b26d for {dataset_id} Nov 14, 2022 2:08:34 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorageSAS is : 201 When user adds pipeline using AzureBlobStorageSAS with dataset Parquet config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:08:34 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureBlobStorageSAS Scenario: Azure Blob Storage of SAS with Excel -> test # features/connectors/Azure/AzureBlobStorageSAS.feature:46 Nov 14, 2022 2:08:55 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:08:55 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 3b7280dd-cf5d-44d2-a30c-9a74ccc4a8d9 for {dataset_id} Nov 14, 2022 2:08:55 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureBlobStorageSAS is : 201 When user adds pipeline using AzureBlobStorageSAS with dataset Excel config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:08:56 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test AzureCosmosDB components @component @smoke @AzureCosmosDB Scenario: test -> CosmosDB # features/connectors/Azure/AzureCosmosDB.feature:5 Nov 14, 2022 2:09:17 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:09:17 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:09:17 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: d947d7d5-2179-4494-b1b6-043a098e0fca for {dataset_id} Nov 14, 2022 2:09:17 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureCosmosDB is : 201 When user adds pipeline using AzureCosmosDB without dataset config for output component with Update config # AddPipelineStep.addPipelineWithSetting(String,String,String) Nov 14, 2022 2:09:17 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureCosmosDB Scenario: CosmosDB -> test # features/connectors/Azure/AzureCosmosDB.feature:11 Nov 14, 2022 2:09:39 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:09:39 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: d947d7d5-2179-4494-b1b6-043a098e0fca for {dataset_id} Nov 14, 2022 2:09:39 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureCosmosDB is : 201 When user adds pipeline using AzureCosmosDB without dataset config for input component without config # AddPipelineStep.addPipeline(String,String) Nov 14, 2022 2:09:39 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test AzureDataLakeStorageGen2 components @component @AzureDataLakeStorageGen2 Scenario: test -> Azure Data Lake Storage Gen2 with CSV # features/connectors/Azure/AzureDataLakeStorageGen2.feature:5 Nov 14, 2022 2:09:54 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:09:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:09:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 1dbb9798-715e-4420-bcfe-08a4ce54cd3a for {dataset_id} Nov 14, 2022 2:09:54 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2 is : 201 When user adds pipeline using AzureDataLakeStorageGen2 with dataset Csv config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:09:55 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureDataLakeStorageGen2 Scenario: test -> Azure Data Lake Storage Gen2 other settings with CSV # features/connectors/Azure/AzureDataLakeStorageGen2.feature:11 Nov 14, 2022 2:10:16 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 9a5b7ef4-d543-4090-ada3-5084d3e75ccf for {os-test_id} Nov 14, 2022 2:10:16 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 2f3f5524-6600-46f6-b808-d0772a0b8109 for {dataset_id} Nov 14, 2022 2:10:16 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2 is : 201 When user adds pipeline using AzureDataLakeStorageGen2 with dataset OtherSettings-Csv config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:10:16 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureDataLakeStorageGen2 Scenario: test -> Azure Data Lake Storage Gen2 Tab settings with CSV # features/connectors/Azure/AzureDataLakeStorageGen2.feature:17 Nov 14, 2022 2:10:38 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:10:38 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:10:38 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: c13b0e0c-b5d3-4cab-a745-41b3353dcd8e for {dataset_id} Nov 14, 2022 2:10:38 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2 is : 201 When user adds pipeline using AzureDataLakeStorageGen2 with dataset Tab-Csv config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:10:38 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureDataLakeStorageGen2 Scenario: test -> Azure Data Lake Storage Gen2 with Avro # features/connectors/Azure/AzureDataLakeStorageGen2.feature:23 Nov 14, 2022 2:10:59 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:10:59 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:10:59 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 48e570dd-d031-4492-a01d-c61c1384c09c for {dataset_id} Nov 14, 2022 2:11:00 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2 is : 201 When user adds pipeline using AzureDataLakeStorageGen2 with dataset Avro config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:11:00 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureDataLakeStorageGen2 Scenario: test -> Azure Data Lake Storage Gen2 with Json # features/connectors/Azure/AzureDataLakeStorageGen2.feature:29 Nov 14, 2022 2:11:21 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:11:21 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:11:21 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 47f1b493-320c-455e-b604-d1d726394e62 for {dataset_id} Nov 14, 2022 2:11:22 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2 is : 201 When user adds pipeline using AzureDataLakeStorageGen2 with dataset Json config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:11:22 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureDataLakeStorageGen2 Scenario: test -> Azure Data Lake Storage Gen2 with Parquet # features/connectors/Azure/AzureDataLakeStorageGen2.feature:35 Nov 14, 2022 2:11:43 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:11:43 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:11:43 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 855396f8-a1fd-4faf-a089-d82cd44dadea for {dataset_id} Nov 14, 2022 2:11:43 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2 is : 201 When user adds pipeline using AzureDataLakeStorageGen2 with dataset Parquet config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:11:43 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureDataLakeStorageGen2 Scenario: Azure Data Lake Storage Gen2 with CSV -> test # features/connectors/Azure/AzureDataLakeStorageGen2.feature:41 Nov 14, 2022 2:12:05 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:12:05 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 1dbb9798-715e-4420-bcfe-08a4ce54cd3a for {dataset_id} Nov 14, 2022 2:12:05 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2 is : 201 When user adds pipeline using AzureDataLakeStorageGen2 with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:12:05 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureDataLakeStorageGen2 Scenario: Azure Data Lake Storage Gen2 other settings with CSV -> test # features/connectors/Azure/AzureDataLakeStorageGen2.feature:47 Nov 14, 2022 2:12:26 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:12:26 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 2f3f5524-6600-46f6-b808-d0772a0b8109 for {dataset_id} Nov 14, 2022 2:12:27 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2 is : 201 When user adds pipeline using AzureDataLakeStorageGen2 with dataset OtherSettings-Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:12:27 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureDataLakeStorageGen2 Scenario: Azure Data Lake Storage Gen2 Tab settings with CSV -> test # features/connectors/Azure/AzureDataLakeStorageGen2.feature:53 Nov 14, 2022 2:12:48 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:12:48 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: c13b0e0c-b5d3-4cab-a745-41b3353dcd8e for {dataset_id} Nov 14, 2022 2:12:48 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2 is : 201 When user adds pipeline using AzureDataLakeStorageGen2 with dataset Tab-Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:12:48 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureDataLakeStorageGen2 Scenario: Azure Data Lake Storage Gen2 with Avro -> test # features/connectors/Azure/AzureDataLakeStorageGen2.feature:59 Nov 14, 2022 2:13:10 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:13:10 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 48e570dd-d031-4492-a01d-c61c1384c09c for {dataset_id} Nov 14, 2022 2:13:10 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2 is : 201 When user adds pipeline using AzureDataLakeStorageGen2 with dataset Avro config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:13:10 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureDataLakeStorageGen2 Scenario: Azure Data Lake Storage Gen2 with Json -> test # features/connectors/Azure/AzureDataLakeStorageGen2.feature:65 Nov 14, 2022 2:13:31 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:13:31 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 47f1b493-320c-455e-b604-d1d726394e62 for {dataset_id} Nov 14, 2022 2:13:32 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2 is : 201 When user adds pipeline using AzureDataLakeStorageGen2 with dataset Json config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:13:32 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureDataLakeStorageGen2 Scenario: Azure Data Lake Storage Gen2 with Parquet -> test # features/connectors/Azure/AzureDataLakeStorageGen2.feature:71 Nov 14, 2022 2:13:53 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:13:53 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 855396f8-a1fd-4faf-a089-d82cd44dadea for {dataset_id} Nov 14, 2022 2:13:53 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2 is : 201 When user adds pipeline using AzureDataLakeStorageGen2 with dataset Parquet config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:13:53 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test AzureDataLakeStorageGen2SAS components @component @AzureDataLakeStorageGen2SAS Scenario: test -> Azure Data Lake Storage Gen2 of SAS with CSV # features/connectors/Azure/AzureDataLakeStorageGen2SAS.feature:5 Nov 14, 2022 2:14:15 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:14:15 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:14:15 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: d15fb577-5e02-4b81-bc86-2ca30b01ae55 for {dataset_id} Nov 14, 2022 2:14:15 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2SAS is : 201 When user adds pipeline using AzureDataLakeStorageGen2SAS with dataset Csv config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:14:15 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureDataLakeStorageGen2SAS Scenario: test -> Azure Data Lake Storage Gen2 of SAS with Avro # features/connectors/Azure/AzureDataLakeStorageGen2SAS.feature:11 Nov 14, 2022 2:14:36 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:14:36 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:14:36 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 6a97dfa7-30fa-4811-90a7-ecd3139371c5 for {dataset_id} Nov 14, 2022 2:14:37 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2SAS is : 201 When user adds pipeline using AzureDataLakeStorageGen2SAS with dataset Avro config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:14:37 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureDataLakeStorageGen2SAS Scenario: test -> Azure Data Lake Storage Gen2 of SAS with Json # features/connectors/Azure/AzureDataLakeStorageGen2SAS.feature:17 Nov 14, 2022 2:14:58 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:14:58 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:14:58 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: dd247585-7502-422b-9859-ead22c3e5788 for {dataset_id} Nov 14, 2022 2:14:58 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2SAS is : 201 When user adds pipeline using AzureDataLakeStorageGen2SAS with dataset Json config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:14:58 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureDataLakeStorageGen2SAS Scenario: test -> Azure Data Lake Storage Gen2 of SAS with Parquet # features/connectors/Azure/AzureDataLakeStorageGen2SAS.feature:23 Nov 14, 2022 2:15:20 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:15:20 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:15:20 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 94926297-4b4f-4dcc-9ad8-5c8247966e11 for {dataset_id} Nov 14, 2022 2:15:20 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2SAS is : 201 When user adds pipeline using AzureDataLakeStorageGen2SAS with dataset Parquet config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:15:20 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureDataLakeStorageGen2SAS Scenario: Azure Data Lake Storage Gen2 of SAS with CSV -> test # features/connectors/Azure/AzureDataLakeStorageGen2SAS.feature:29 Nov 14, 2022 2:15:42 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:15:42 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: d15fb577-5e02-4b81-bc86-2ca30b01ae55 for {dataset_id} Nov 14, 2022 2:15:42 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2SAS is : 201 When user adds pipeline using AzureDataLakeStorageGen2SAS with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:15:42 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureDataLakeStorageGen2SAS Scenario: Azure Data Lake Storage Gen2 of SAS with Avro -> test # features/connectors/Azure/AzureDataLakeStorageGen2SAS.feature:36 Nov 14, 2022 2:16:03 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:16:03 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 6a97dfa7-30fa-4811-90a7-ecd3139371c5 for {dataset_id} Nov 14, 2022 2:16:03 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2SAS is : 201 When user adds pipeline using AzureDataLakeStorageGen2SAS with dataset Avro config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:16:04 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @AzureDataLakeStorageGen2SAS Scenario: Azure Data Lake Storage Gen2 of SAS with Json -> test # features/connectors/Azure/AzureDataLakeStorageGen2SAS.feature:42 Nov 14, 2022 2:16:25 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:16:25 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: dd247585-7502-422b-9859-ead22c3e5788 for {dataset_id} Nov 14, 2022 2:16:25 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2SAS is : 201 When user adds pipeline using AzureDataLakeStorageGen2SAS with dataset Json config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:16:25 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureDataLakeStorageGen2SAS Scenario: Azure Data Lake Storage Gen2 of SAS with Parquet -> test # features/connectors/Azure/AzureDataLakeStorageGen2SAS.feature:48 Nov 14, 2022 2:16:47 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:16:47 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 94926297-4b4f-4dcc-9ad8-5c8247966e11 for {dataset_id} Nov 14, 2022 2:16:47 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureDataLakeStorageGen2SAS is : 201 When user adds pipeline using AzureDataLakeStorageGen2SAS with dataset Parquet config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:16:47 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test AzureEventHubs components @component @smoke @AzureEventHubs Scenario: test -> Azure Event Hubs # features/connectors/Azure/AzureEventHubs.feature:5 Nov 14, 2022 2:17:08 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:17:08 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:17:08 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: ef8b7e1a-e609-4244-b446-ef2cf95c0a19 for {dataset_id} Nov 14, 2022 2:17:09 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureEventHubs is : 201 When user adds pipeline using AzureEventHubs with dataset Csv config for output component with RoundRobin config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:17:09 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureEventHubs Scenario: Azure Event Hubs -> test # features/connectors/Azure/AzureEventHubs.feature:11 Nov 14, 2022 2:17:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:17:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: ef8b7e1a-e609-4244-b446-ef2cf95c0a19 for {dataset_id} Nov 14, 2022 2:17:30 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureEventHubs is : 201 When user adds pipeline using AzureEventHubs with dataset Csv config for input component with Earliest config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:17:30 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 2:19:04 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component @smoke @AzureEventHubs Scenario: test -> Azure Event Hubs with AVRO # features/connectors/Azure/AzureEventHubs.feature:18 Nov 14, 2022 2:19:07 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 6f8cea99-dbe0-4232-8258-8a694f366a4f for {os-test_id} Nov 14, 2022 2:19:07 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: f53c0e23-7d3f-4741-8c6b-fd2089dbcb36 for {dataset_id} Nov 14, 2022 2:19:07 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureEventHubs is : 201 When user adds pipeline using AzureEventHubs with dataset Avro config for output component with RoundRobin config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:19:07 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureEventHubs Scenario: Azure Event Hubs with AVRO -> test # features/connectors/Azure/AzureEventHubs.feature:24 Nov 14, 2022 2:19:28 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:19:28 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: f53c0e23-7d3f-4741-8c6b-fd2089dbcb36 for {dataset_id} Nov 14, 2022 2:19:29 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureEventHubs is : 201 When user adds pipeline using AzureEventHubs with dataset Avro config for input component with Earliest config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:19:29 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 2:21:02 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component @smoke @AzureEventHubs Scenario: test -> Azure Event Hubs with JSON # features/connectors/Azure/AzureEventHubs.feature:31 Nov 14, 2022 2:21:05 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:21:05 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:21:05 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 2498f3cf-d2d0-4870-882b-76bde95f6c79 for {dataset_id} Nov 14, 2022 2:21:05 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureEventHubs is : 201 When user adds pipeline using AzureEventHubs with dataset Json config for output component with RoundRobin config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:21:06 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureEventHubs Scenario: Azure Event Hubs with JSON -> test # features/connectors/Azure/AzureEventHubs.feature:37 Nov 14, 2022 2:21:27 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:21:27 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 2498f3cf-d2d0-4870-882b-76bde95f6c79 for {dataset_id} Nov 14, 2022 2:21:28 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureEventHubs is : 201 When user adds pipeline using AzureEventHubs with dataset Json config for input component with Earliest config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:21:28 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 2:23:04 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component @smoke @AzureEventHubs Scenario: test -> Azure Event Hubs with TEXT # features/connectors/Azure/AzureEventHubs.feature:44 Nov 14, 2022 2:23:07 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 010faa41-91b2-4695-8824-ef6c7d22041d for {os-test_id} Nov 14, 2022 2:23:07 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: edd2ab1a-e1c2-40f8-97c3-39e6da14fe58 for {dataset_id} Nov 14, 2022 2:23:07 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureEventHubs is : 201 When user adds pipeline using AzureEventHubs with dataset Text config for output component with RoundRobin config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:23:07 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @AzureEventHubs Scenario: Azure Event Hubs with TEXT -> test # features/connectors/Azure/AzureEventHubs.feature:50 Nov 14, 2022 2:23:29 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:23:29 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: edd2ab1a-e1c2-40f8-97c3-39e6da14fe58 for {dataset_id} Nov 14, 2022 2:23:29 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureEventHubs is : 201 When user adds pipeline using AzureEventHubs with dataset Text config for input component with Earliest config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:23:29 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 2:25:02 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component @smoke @AzureEventHubs Scenario: Azure Event Hubs with AVRO -> test # features/connectors/Azure/AzureEventHubs.feature:58 Nov 14, 2022 2:25:05 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:25:05 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: f53c0e23-7d3f-4741-8c6b-fd2089dbcb36 for {dataset_id} Nov 14, 2022 2:25:05 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureEventHubs is : 201 When user adds pipeline using AzureEventHubs with dataset Avro config for input component with Earliest_AzureSAS config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:25:05 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 2:26:39 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component @Box Feature: test Box components - Enterprise ID @component @Box @smoke @cleanUpBox Scenario: test data -> Box with CSV # features/connectors/Box/Box_Enterprise_Id.feature:6 Box folders content is being removed - CleanUpHook Nov 14, 2022 2:26:44 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:26:44 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:26:44 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: b78a73d1-245c-4eb0-8cb9-5fca30dcb122 for {dataset_id} Nov 14, 2022 2:26:44 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_Enterprise_Id is : 201 When user adds pipeline using Box_Enterprise_Id with dataset Csv config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:26:44 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box @smoke Scenario: Box with CSV -> test # features/connectors/Box/Box_Enterprise_Id.feature:13 Nov 14, 2022 2:27:24 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:27:24 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: b78a73d1-245c-4eb0-8cb9-5fca30dcb122 for {dataset_id} Nov 14, 2022 2:27:24 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_Enterprise_Id is : 201 When user adds pipeline using Box_Enterprise_Id with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:27:24 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: test data -> Box with Avro # features/connectors/Box/Box_Enterprise_Id.feature:25 Nov 14, 2022 2:28:04 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:28:04 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:28:04 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 9d130e84-baf8-4c59-b5fe-12e91e2add2a for {dataset_id} Nov 14, 2022 2:28:04 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_Enterprise_Id is : 201 When user adds pipeline using Box_Enterprise_Id with dataset Avro config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:28:04 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: Box with Avro -> test # features/connectors/Box/Box_Enterprise_Id.feature:31 Nov 14, 2022 2:28:44 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:28:44 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 9d130e84-baf8-4c59-b5fe-12e91e2add2a for {dataset_id} Nov 14, 2022 2:28:44 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_Enterprise_Id is : 201 When user adds pipeline using Box_Enterprise_Id with dataset Avro config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:28:44 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: test data -> Box with Json # features/connectors/Box/Box_Enterprise_Id.feature:37 Nov 14, 2022 2:29:24 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:29:24 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:29:24 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 68af078d-fd30-46ed-9d37-71ebee74f28a for {dataset_id} Nov 14, 2022 2:29:24 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_Enterprise_Id is : 201 When user adds pipeline using Box_Enterprise_Id with dataset Json config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:29:24 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: Box with Json -> test # features/connectors/Box/Box_Enterprise_Id.feature:43 Nov 14, 2022 2:30:04 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:30:04 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 68af078d-fd30-46ed-9d37-71ebee74f28a for {dataset_id} Nov 14, 2022 2:30:04 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_Enterprise_Id is : 201 When user adds pipeline using Box_Enterprise_Id with dataset Json config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:30:04 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: test data -> Box with Excel # features/connectors/Box/Box_Enterprise_Id.feature:49 Nov 14, 2022 2:30:44 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:30:44 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:30:44 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 2e76e3c4-a437-4cc1-b57a-5bf4f4163629 for {dataset_id} Nov 14, 2022 2:30:44 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_Enterprise_Id is : 201 When user adds pipeline using Box_Enterprise_Id with dataset Excel config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:30:44 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: Box with Excel -> test # features/connectors/Box/Box_Enterprise_Id.feature:55 Nov 14, 2022 2:31:24 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:31:24 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 2e76e3c4-a437-4cc1-b57a-5bf4f4163629 for {dataset_id} Nov 14, 2022 2:31:24 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_Enterprise_Id is : 201 When user adds pipeline using Box_Enterprise_Id with dataset Excel config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:31:24 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: Box with Excel (HTML) -> test # features/connectors/Box/Box_Enterprise_Id.feature:61 Nov 14, 2022 2:32:04 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:32:04 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 3306aa60-8df5-42e2-acba-fca24864d06d for {dataset_id} Nov 14, 2022 2:32:05 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_Enterprise_Id is : 201 When user adds pipeline using Box_Enterprise_Id with dataset Excel-HTML-Read config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:32:05 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Feature: test Box components - User ID @component @Box @cleanUpBox Scenario: test data -> Box with CSV # features/connectors/Box/Box_User_Id.feature:6 Box folders content is being removed - CleanUpHook Nov 14, 2022 2:32:50 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:32:50 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:32:50 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 9226dd1a-0d92-4df5-bfe2-fdfad6cfd2f9 for {dataset_id} Nov 14, 2022 2:32:50 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_User_Id is : 201 When user adds pipeline using Box_User_Id with dataset Csv config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:32:50 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: Box with CSV -> test # features/connectors/Box/Box_User_Id.feature:12 Nov 14, 2022 2:33:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:33:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 9226dd1a-0d92-4df5-bfe2-fdfad6cfd2f9 for {dataset_id} Nov 14, 2022 2:33:30 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_User_Id is : 201 When user adds pipeline using Box_User_Id with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:33:30 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box @smoke Scenario: test data -> Box with Avro # features/connectors/Box/Box_User_Id.feature:25 Nov 14, 2022 2:34:10 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:34:10 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:34:10 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 54b15c85-93d2-4677-b22a-1bec798cc9f1 for {dataset_id} Nov 14, 2022 2:34:10 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_User_Id is : 201 When user adds pipeline using Box_User_Id with dataset Avro config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:34:10 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box @smoke Scenario: Box with Avro -> test # features/connectors/Box/Box_User_Id.feature:32 Nov 14, 2022 2:34:50 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:34:50 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 54b15c85-93d2-4677-b22a-1bec798cc9f1 for {dataset_id} Nov 14, 2022 2:34:50 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_User_Id is : 201 When user adds pipeline using Box_User_Id with dataset Avro config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:34:50 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: test data -> Box with Json # features/connectors/Box/Box_User_Id.feature:38 Nov 14, 2022 2:35:30 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:35:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:35:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 2c3cbeb3-a2f3-40ab-9a25-1217c19fbf46 for {dataset_id} Nov 14, 2022 2:35:30 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_User_Id is : 201 When user adds pipeline using Box_User_Id with dataset Json config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:35:30 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: Box with Json -> test # features/connectors/Box/Box_User_Id.feature:44 Nov 14, 2022 2:36:10 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:36:10 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 2c3cbeb3-a2f3-40ab-9a25-1217c19fbf46 for {dataset_id} Nov 14, 2022 2:36:10 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_User_Id is : 201 When user adds pipeline using Box_User_Id with dataset Json config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:36:10 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: test data -> Box with Excel # features/connectors/Box/Box_User_Id.feature:50 Nov 14, 2022 2:36:50 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:36:50 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:36:50 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 506de0dd-423d-440f-ae7f-6ccba2642787 for {dataset_id} Nov 14, 2022 2:36:50 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_User_Id is : 201 When user adds pipeline using Box_User_Id with dataset Excel config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:36:50 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: Box with Excel -> test # features/connectors/Box/Box_User_Id.feature:56 Nov 14, 2022 2:37:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:37:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 506de0dd-423d-440f-ae7f-6ccba2642787 for {dataset_id} Nov 14, 2022 2:37:30 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_User_Id is : 201 When user adds pipeline using Box_User_Id with dataset Excel config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:37:30 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Box Scenario: Box with Excel (HTML) -> test # features/connectors/Box/Box_User_Id.feature:62 Nov 14, 2022 2:38:10 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:38:10 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 61e406f1-ea71-4ce0-9040-07865c02157c for {dataset_id} Nov 14, 2022 2:38:10 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Box_User_Id is : 201 When user adds pipeline using Box_User_Id with dataset Excel-HTML-Read config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:38:10 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test Couchbase components @component @Couchbase Scenario: Couchbase -> test # features/connectors/Couchbase.feature:5 Nov 14, 2022 2:38:50 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:38:50 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: edfce939-11b9-431e-a571-58bc18c4aa86 for {dataset_id} Nov 14, 2022 2:38:50 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Couchbase is : 201 When user adds pipeline using Couchbase with dataset JSON config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:38:50 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Couchbase Scenario: test -> Couchbase # features/connectors/Couchbase.feature:11 Nov 14, 2022 2:39:12 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:39:12 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:39:12 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: edfce939-11b9-431e-a571-58bc18c4aa86 for {dataset_id} Nov 14, 2022 2:39:12 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Couchbase is : 201 When user adds pipeline using Couchbase with dataset JSON config for output component with Insert config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:39:12 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database Feature: all Aurora components @component @Database @smoke @Aurora Scenario: test data -> Amazon Aurora # features/connectors/Database/Aurora.feature:5 Nov 14, 2022 2:39:34 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:39:34 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:39:34 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: eae5220b-0c5a-4af8-8b59-a55492677996 for {dataset_id} Nov 14, 2022 2:39:34 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Aurora is : 201 When user adds database pipeline using Aurora with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:39:34 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @Aurora Scenario: Amazon Aurora -> test # features/connectors/Database/Aurora.feature:11 Nov 14, 2022 2:39:55 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:39:55 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: eae5220b-0c5a-4af8-8b59-a55492677996 for {dataset_id} Nov 14, 2022 2:39:55 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Aurora is : 201 When user adds database pipeline using Aurora with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:39:55 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database Feature: test AzureSynabase components @component @Database @AzureSynapse Scenario: test data -> Azure Synabase with bulkload # features/connectors/Database/AzureSynabase.feature:5 Nov 14, 2022 2:40:11 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:40:11 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:40:11 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 84ce7566-d2dd-4f6c-a2b6-01bcf6ab9a88 for {dataset_id} Nov 14, 2022 2:40:11 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureSynapse is : 201 When user adds database pipeline using AzureSynapse with dataset Table config for output component with Bulkload-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:40:11 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @AzureSynapse Scenario: test data -> Azure Synabase # features/connectors/Database/AzureSynabase.feature:11 Nov 14, 2022 2:40:32 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:40:32 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:40:32 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 84ce7566-d2dd-4f6c-a2b6-01bcf6ab9a88 for {dataset_id} Nov 14, 2022 2:40:33 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureSynapse is : 201 When user adds database pipeline using AzureSynapse with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:40:33 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @AzureSynapse Scenario: Azure Synabase -> test # features/connectors/Database/AzureSynabase.feature:17 Nov 14, 2022 2:40:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:40:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 84ce7566-d2dd-4f6c-a2b6-01bcf6ab9a88 for {dataset_id} Nov 14, 2022 2:40:54 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using AzureSynapse is : 201 When user adds database pipeline using AzureSynapse with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:40:54 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database Feature: test DeltaLake components @component @Database @smoke @DeltaLake Scenario: test data -> DeltaLake # features/connectors/Database/DeltaLake.feature:5 Nov 14, 2022 2:43:54 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:43:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:43:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 55835cee-8ec2-409a-a662-b3f8b052551f for {dataset_id} Nov 14, 2022 2:43:55 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using DeltaLake is : 201 When user adds database pipeline using DeltaLake with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:43:55 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Nov 14, 2022 2:44:41 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 Nov 14, 2022 2:45:27 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 Nov 14, 2022 2:46:13 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() java.lang.AssertionError: The pipeline does not run successfully, the last run status is [error, (com.simba.spark.support.exceptions.ErrorException) [Simba][SparkJDBCDriver](700100) Connection timeout expired. Details: None.] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkPipelineStatus(TPDAPIOperation.java:431) at steps.AddPipelineStep.checkPipelineStatus(AddPipelineStep.java:83) at ✽.the lastExecutionStatus for the pipeline is "finished"(features/connectors/Database/DeltaLake.feature:8) @component @Database @smoke @DeltaLake Scenario: DeltaLake -> test # features/connectors/Database/DeltaLake.feature:11 Nov 14, 2022 2:46:59 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:46:59 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 55835cee-8ec2-409a-a662-b3f8b052551f for {dataset_id} Nov 14, 2022 2:46:59 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using DeltaLake is : 201 When user adds database pipeline using DeltaLake with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:46:59 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Nov 14, 2022 2:47:45 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database Feature: test Derby components @component @Database @smoke @Derby Scenario: test data -> Derby # features/connectors/Database/Derby.feature:5 Nov 14, 2022 2:48:43 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:48:43 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:48:43 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: e71fbca7-ac94-4efa-9285-d6c501ee9497 for {dataset_id} Nov 14, 2022 2:48:43 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Derby is : 201 When user adds database pipeline using Derby with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:48:44 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @Derby Scenario: Derby -> test # features/connectors/Database/Derby.feature:11 Nov 14, 2022 2:48:59 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:48:59 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: e71fbca7-ac94-4efa-9285-d6c501ee9497 for {dataset_id} Nov 14, 2022 2:48:59 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Derby is : 201 When user adds database pipeline using Derby with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:49:00 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database Feature: test Mssql components @component @Database @smoke @MSSQL Scenario: test data -> Amazon MSSQL # features/connectors/Database/MSSQL.feature:5 Nov 14, 2022 2:49:15 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:49:15 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:49:15 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a031e1ea-34c3-40c0-99f2-de0881932f5c for {dataset_id} Nov 14, 2022 2:49:15 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MSSQL is : 201 When user adds database pipeline using MSSQL with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:49:16 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @MSSQL Scenario: test data -> Amazon MSSQL with the authentication type of basic # features/connectors/Database/MSSQL.feature:11 Nov 14, 2022 2:49:37 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:49:37 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:49:37 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 27f16574-0d02-4254-8a53-e8c8697c52e3 for {dataset_id} Nov 14, 2022 2:49:37 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MSSQLBasic is : 201 When user adds database pipeline using MSSQLBasic with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:49:37 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @MSSQL Scenario: Amazon MSSQL -> test # features/connectors/Database/MSSQL.feature:17 Nov 14, 2022 2:49:53 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:49:53 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a031e1ea-34c3-40c0-99f2-de0881932f5c for {dataset_id} Nov 14, 2022 2:49:53 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MSSQL is : 201 When user adds database pipeline using MSSQL with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:49:53 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @MSSQL Scenario: Amazon MSSQL with the authentication type of basic -> test # features/connectors/Database/MSSQL.feature:23 Nov 14, 2022 2:50:15 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:50:15 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 27f16574-0d02-4254-8a53-e8c8697c52e3 for {dataset_id} Nov 14, 2022 2:50:15 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MSSQLBasic is : 201 When user adds database pipeline using MSSQLBasic with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:50:15 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database Feature: test MariaDB components @component @Database @smoke @MariaDB Scenario: test data -> Amazon MariaDB # features/connectors/Database/MariaDB.feature:5 Nov 14, 2022 2:50:31 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:50:31 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:50:31 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 54947e5f-5ea8-4b17-a3ab-b90815441b3e for {dataset_id} Nov 14, 2022 2:50:31 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MariaDB is : 201 When user adds database pipeline using MariaDB with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:50:31 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @MariaDB Scenario: test data -> Amazon MariaDB with the authentication type of basic # features/connectors/Database/MariaDB.feature:11 Nov 14, 2022 2:50:46 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:50:46 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:50:46 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 95f2bf50-4869-45dd-bf2a-700bc2ed7e34 for {dataset_id} Nov 14, 2022 2:50:47 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MariaDBBasic is : 201 When user adds database pipeline using MariaDBBasic with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:50:47 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @MariaDB Scenario: Amazon MariaDB -> test # features/connectors/Database/MariaDB.feature:17 Nov 14, 2022 2:51:02 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:51:02 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 54947e5f-5ea8-4b17-a3ab-b90815441b3e for {dataset_id} Nov 14, 2022 2:51:02 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MariaDB is : 201 When user adds database pipeline using MariaDB with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:51:03 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @MariaDB Scenario: Amazon MariaDB with the authentication type of basic-> test # features/connectors/Database/MariaDB.feature:23 Nov 14, 2022 2:51:18 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:51:18 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 95f2bf50-4869-45dd-bf2a-700bc2ed7e34 for {dataset_id} Nov 14, 2022 2:51:18 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MariaDBBasic is : 201 When user adds database pipeline using MariaDBBasic with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:51:19 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database Feature: test Mysql components @component @Database @smoke @MySql Scenario: test data -> Amazon Mysql # features/connectors/Database/MySql.feature:5 Nov 14, 2022 2:51:34 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:51:34 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:51:34 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 3bdd2b81-8e8f-4497-9703-1df7680d2063 for {dataset_id} Nov 14, 2022 2:51:34 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MySQL is : 201 When user adds database pipeline using MySQL with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:51:34 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @MySql Scenario: test data -> Amazon Mysql with the authentication type of basic # features/connectors/Database/MySql.feature:11 Nov 14, 2022 2:51:50 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:51:50 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:51:50 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 633ea9fd-aeb1-424c-961d-644021e936c5 for {dataset_id} Nov 14, 2022 2:51:50 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MySQLBasic is : 201 When user adds database pipeline using MySQLBasic with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:51:50 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @MySql Scenario: Amazon MySQL -> test # features/connectors/Database/MySql.feature:17 Nov 14, 2022 2:52:06 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:52:06 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 3bdd2b81-8e8f-4497-9703-1df7680d2063 for {dataset_id} Nov 14, 2022 2:52:06 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MySQL is : 201 When user adds database pipeline using MySQL with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:52:06 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @MySql Scenario: Amazon Mysql with the authentication type of basic -> test # features/connectors/Database/MySql.feature:23 Nov 14, 2022 2:52:22 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:52:22 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 633ea9fd-aeb1-424c-961d-644021e936c5 for {dataset_id} Nov 14, 2022 2:52:22 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MySQLBasic is : 201 When user adds database pipeline using MySQLBasic with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:52:22 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database Feature: test PostgreSQL components @component @Database @smoke @PostgreSQL Scenario: test data -> Amazon PostgreSQL # features/connectors/Database/PostgreSQL.feature:5 Nov 14, 2022 2:52:37 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:52:37 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:52:37 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: ae0b3dee-4116-4cae-8d5a-4d167b9b57f4 for {dataset_id} Nov 14, 2022 2:52:38 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using PostgreSQL is : 201 When user adds database pipeline using PostgreSQL with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:52:38 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @PostgreSQL Scenario: test data -> Amazon PostgreSQL with the authentication type of basic # features/connectors/Database/PostgreSQL.feature:11 Nov 14, 2022 2:52:53 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:52:53 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:52:53 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: fe323d3a-d9be-4f75-92c0-dcade672e937 for {dataset_id} Nov 14, 2022 2:52:53 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using PostgreSQLBasic is : 201 When user adds database pipeline using PostgreSQLBasic with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:52:54 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @PostgreSQL Scenario: Amazon PostgreSQL -> test # features/connectors/Database/PostgreSQL.feature:17 Nov 14, 2022 2:53:09 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:53:09 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: ae0b3dee-4116-4cae-8d5a-4d167b9b57f4 for {dataset_id} Nov 14, 2022 2:53:09 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using PostgreSQL is : 201 When user adds database pipeline using PostgreSQL with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:53:09 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @PostgreSQL Scenario: Amazon PostgreSQL with the authentication type of basic -> test # features/connectors/Database/PostgreSQL.feature:23 Nov 14, 2022 2:53:25 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:53:25 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: fe323d3a-d9be-4f75-92c0-dcade672e937 for {dataset_id} Nov 14, 2022 2:53:25 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using PostgreSQLBasic is : 201 When user adds database pipeline using PostgreSQLBasic with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:53:25 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database Feature: test RedShift components @component @Database @smoke @RedShift Scenario: test data -> Amazon RedShift # features/connectors/Database/RedShift.feature:5 Nov 14, 2022 2:53:41 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:53:41 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:53:41 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 01d341e0-d16b-4ee5-96ad-ea9864fde591 for {dataset_id} Nov 14, 2022 2:53:41 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Redshift is : 201 When user adds database pipeline using Redshift with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:53:41 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @RedShift Scenario: test data -> Amazon RedShift with bulkload # features/connectors/Database/RedShift.feature:11 Nov 14, 2022 2:53:56 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:53:56 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:53:56 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 01d341e0-d16b-4ee5-96ad-ea9864fde591 for {dataset_id} Nov 14, 2022 2:53:57 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Redshift is : 201 When user adds database pipeline using Redshift with dataset Table config for output component with Bulkload-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:53:57 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @RedShift Scenario: Amazon RedShift -> test # features/connectors/Database/RedShift.feature:17 Nov 14, 2022 2:55:08 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:55:08 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 01d341e0-d16b-4ee5-96ad-ea9864fde591 for {dataset_id} Nov 14, 2022 2:55:08 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Redshift is : 201 When user adds database pipeline using Redshift with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:55:08 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database Feature: test SingleStore components @component @Database @SingleStore Scenario: SingleStore -> test # features/connectors/Database/SingleStore.feature:5 Nov 14, 2022 2:55:24 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:55:24 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 6f6625ee-c120-4807-94bd-af083f8a28c7 for {dataset_id} Nov 14, 2022 2:55:24 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using SingleStore is : 201 When user adds database pipeline using SingleStore with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:55:24 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @SingleStore Scenario: test data -> SingleStore # features/connectors/Database/SingleStore.feature:11 Nov 14, 2022 2:55:40 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:55:40 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:55:40 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 6f6625ee-c120-4807-94bd-af083f8a28c7 for {dataset_id} Nov 14, 2022 2:55:40 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using SingleStore is : 201 When user adds database pipeline using SingleStore with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:55:40 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database Feature: test Snowflake components @component @Database @smoke @Snowflake Scenario: test data -> Snowflake # features/connectors/Database/Snowflake.feature:5 Nov 14, 2022 2:55:55 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:55:55 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:55:55 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: f9c114da-96b8-4175-8fba-0f2703d91ac8 for {dataset_id} Nov 14, 2022 2:55:57 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Snowflake is : 201 When user adds database pipeline using Snowflake with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:55:57 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @smoke @Snowflake Scenario: Snowflake -> test # features/connectors/Database/Snowflake.feature:11 Nov 14, 2022 2:56:19 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:56:19 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: f9c114da-96b8-4175-8fba-0f2703d91ac8 for {dataset_id} Nov 14, 2022 2:56:20 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Snowflake is : 201 When user adds database pipeline using Snowflake with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:56:20 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @Snowflake @smoke Scenario: test data -> Snowflake with the authentication type of basic # features/connectors/Database/Snowflake.feature:17 Nov 14, 2022 2:56:42 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:56:42 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:56:42 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: b7acc2d9-4877-49d0-b62e-de2c28aad6de for {dataset_id} Nov 14, 2022 2:56:42 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using SnowflakeBasic is : 201 When user adds database pipeline using SnowflakeBasic with dataset Table config for output component with Insert-CreateTable config # AddPipelineStep.addDatabasePipeline(String,String,String,String) Nov 14, 2022 2:56:42 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Database @Snowflake @smoke Scenario: Snowflake with the authentication type of basic -> test # features/connectors/Database/Snowflake.feature:23 Nov 14, 2022 2:57:04 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:57:04 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: b7acc2d9-4877-49d0-b62e-de2c28aad6de for {dataset_id} Nov 14, 2022 2:57:04 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using SnowflakeBasic is : 201 When user adds database pipeline using SnowflakeBasic with dataset Table config for input component without config # AddPipelineStep.addDatabasePipeline(String,String,String) Nov 14, 2022 2:57:04 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test ElasticSearch component @component @smoke @ElasticSearch Scenario: test data -> elasticsearch # features/connectors/ElasticSearch.feature:5 Nov 14, 2022 2:57:26 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:57:26 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:57:26 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: b51f8059-7f1a-447b-959b-6e739c64c612 for {dataset_id} Nov 14, 2022 2:57:26 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using ElasticSearch is : 201 When user adds pipeline using ElasticSearch without dataset config for output component without config # AddPipelineStep.addPipeline(String,String) Nov 14, 2022 2:57:26 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @ElasticSearch Scenario: elasticsearch -> test # features/connectors/ElasticSearch.feature:11 Nov 14, 2022 2:57:42 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:57:42 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: b51f8059-7f1a-447b-959b-6e739c64c612 for {dataset_id} Nov 14, 2022 2:57:42 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using ElasticSearch is : 201 When user adds pipeline using ElasticSearch without dataset config for input component without config # AddPipelineStep.addPipeline(String,String) Nov 14, 2022 2:57:42 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test FTP components @component @smoke @FTP Scenario: test -> FTP with CSV # features/connectors/FTP.feature:5 Nov 14, 2022 2:57:58 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:57:58 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:57:58 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8daa989f-defc-4fa5-9477-5b1d4dcfde31 for {dataset_id} Nov 14, 2022 2:57:58 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using FTP is : 201 When user adds pipeline using FTP with dataset Csv config for output component with Size config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:57:58 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @FTP Scenario: FTP with CSV -> test # features/connectors/FTP.feature:11 Nov 14, 2022 2:58:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:58:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8daa989f-defc-4fa5-9477-5b1d4dcfde31 for {dataset_id} Nov 14, 2022 2:58:14 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using FTP is : 201 When user adds pipeline using FTP with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:58:14 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test GoogleBigQuery components @component @smoke @GoogleBigQuery Scenario: test data -> bigquery # features/connectors/Google/GoogleBigQuery.feature:5 Nov 14, 2022 2:58:29 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 2:58:29 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:58:29 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 74f29a02-6de6-4c6e-aabb-42efdd2ff380 for {dataset_id} Nov 14, 2022 2:58:30 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleBigQuery is : 201 When user adds pipeline using GoogleBigQuery with dataset Table config for output component with Create config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 2:58:30 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @GoogleBigQuery Scenario: bigquery -> test # features/connectors/Google/GoogleBigQuery.feature:11 Nov 14, 2022 2:58:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:58:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 74f29a02-6de6-4c6e-aabb-42efdd2ff380 for {dataset_id} Nov 14, 2022 2:58:52 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleBigQuery is : 201 When user adds pipeline using GoogleBigQuery with dataset Table config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:58:52 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test GoogleCloudStorage components @component @smoke @GoogleCloudStorage Scenario: test -> Google Cloud Storage with CSV # features/connectors/Google/GoogleCloudStorage.feature:5 Nov 14, 2022 2:59:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 2:59:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 73ca16db-7f58-4cae-a59c-bcf6b77a4c9f for {dataset_id} Nov 14, 2022 2:59:14 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset Csv config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:59:14 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @GoogleCloudStorage Scenario: Google Cloud Storage with CSV -> test # features/connectors/Google/GoogleCloudStorage.feature:11 Nov 14, 2022 2:59:29 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 2:59:29 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 73ca16db-7f58-4cae-a59c-bcf6b77a4c9f for {dataset_id} Nov 14, 2022 2:59:30 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:59:30 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with path # features/connectors/Google/GoogleCloudStorage.feature:24 Nov 14, 2022 2:59:51 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 2:59:51 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 0b24a36a-d255-4531-8a89-cc9959ae8ef1 for {dataset_id} Nov 14, 2022 2:59:52 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset Path config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 2:59:52 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with path -> test # features/connectors/Google/GoogleCloudStorage.feature:31 Nov 14, 2022 3:00:07 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:00:07 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 0b24a36a-d255-4531-8a89-cc9959ae8ef1 for {dataset_id} Nov 14, 2022 3:00:08 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset Path config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:00:08 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with Common IO JSON # features/connectors/Google/GoogleCloudStorage.feature:44 Nov 14, 2022 3:00:36 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 3:00:36 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 0c0baa85-2a2d-48d1-abba-2ecdd60be07f for {dataset_id} Nov 14, 2022 3:00:36 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-JSON config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:00:36 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with Common IO JSON -> test # features/connectors/Google/GoogleCloudStorage.feature:51 Nov 14, 2022 3:00:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:00:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 0c0baa85-2a2d-48d1-abba-2ecdd60be07f for {dataset_id} Nov 14, 2022 3:00:52 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-JSON config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:00:52 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with Common IO AVRO # features/connectors/Google/GoogleCloudStorage.feature:64 Nov 14, 2022 3:01:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 3:01:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 7a03f6ab-d4b2-4e6b-aae0-558285096655 for {dataset_id} Nov 14, 2022 3:01:14 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-AVRO config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:01:14 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with Common IO AVRO -> test # features/connectors/Google/GoogleCloudStorage.feature:71 Nov 14, 2022 3:01:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:01:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 7a03f6ab-d4b2-4e6b-aae0-558285096655 for {dataset_id} Nov 14, 2022 3:01:30 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-AVRO config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:01:30 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with Common IO Excel97 # features/connectors/Google/GoogleCloudStorage.feature:78 Nov 14, 2022 3:01:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 3:01:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: dbdc41ff-29a4-443b-9a5b-fa46fc2e0a77 for {dataset_id} Nov 14, 2022 3:01:52 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-Excel97 config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:01:52 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with Common IO Excel97 -> test # features/connectors/Google/GoogleCloudStorage.feature:85 Nov 14, 2022 3:02:08 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:02:08 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: dbdc41ff-29a4-443b-9a5b-fa46fc2e0a77 for {dataset_id} Nov 14, 2022 3:02:08 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-Excel97 config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:02:08 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with Common IO Excel2007 # features/connectors/Google/GoogleCloudStorage.feature:92 Nov 14, 2022 3:02:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 3:02:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: fa432813-d085-409a-924a-0d66dc564a41 for {dataset_id} Nov 14, 2022 3:02:30 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-Excel2007 config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:02:30 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with Common IO Excel2007 -> test # features/connectors/Google/GoogleCloudStorage.feature:99 Nov 14, 2022 3:02:46 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:02:46 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: fa432813-d085-409a-924a-0d66dc564a41 for {dataset_id} Nov 14, 2022 3:02:46 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-Excel2007 config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:02:46 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with Common IO CSV Linux Type # features/connectors/Google/GoogleCloudStorage.feature:106 Nov 14, 2022 3:03:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 3:03:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a6e38aec-af1a-462e-a1c3-44b81f6a85c1 for {dataset_id} Nov 14, 2022 3:03:14 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-LinuxType config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:03:14 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with Common IO CSV Linux Type -> test # features/connectors/Google/GoogleCloudStorage.feature:113 Nov 14, 2022 3:03:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:03:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a6e38aec-af1a-462e-a1c3-44b81f6a85c1 for {dataset_id} Nov 14, 2022 3:03:30 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-LinuxType config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:03:30 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with Common IO CSV Windows Type # features/connectors/Google/GoogleCloudStorage.feature:120 Nov 14, 2022 3:03:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 3:03:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 6150e865-9f4b-43f0-b7ac-80825fbb5083 for {dataset_id} Nov 14, 2022 3:03:52 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-WindowsType config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:03:52 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with Common IO CSV Windows Type -> test # features/connectors/Google/GoogleCloudStorage.feature:127 Nov 14, 2022 3:04:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:04:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 6150e865-9f4b-43f0-b7ac-80825fbb5083 for {dataset_id} Nov 14, 2022 3:04:14 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-WindowsType config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:04:14 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with Common IO CSV ISO-8859-15 # features/connectors/Google/GoogleCloudStorage.feature:134 Nov 14, 2022 3:04:36 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 3:04:36 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 39d2db1f-30f4-449f-bd74-f0c22ff7fa9f for {dataset_id} Nov 14, 2022 3:04:36 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-ISO-8859-15 config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:04:36 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with Common IO CSV ISO-8859-15 -> test # features/connectors/Google/GoogleCloudStorage.feature:141 Nov 14, 2022 3:04:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:04:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 39d2db1f-30f4-449f-bd74-f0c22ff7fa9f for {dataset_id} Nov 14, 2022 3:04:52 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-ISO-8859-15 config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:04:52 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with Common IO CSV Header # features/connectors/Google/GoogleCloudStorage.feature:148 Nov 14, 2022 3:05:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 3:05:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 212a9cf7-f5f3-46db-ae84-798c6a245b9a for {dataset_id} Nov 14, 2022 3:05:14 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-Header config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:05:14 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with Common IO CSV Header -> test # features/connectors/Google/GoogleCloudStorage.feature:155 Nov 14, 2022 3:05:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:05:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 212a9cf7-f5f3-46db-ae84-798c6a245b9a for {dataset_id} Nov 14, 2022 3:05:30 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-Header config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:05:30 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with Common IO CSV PipeFieldSeparator # features/connectors/Google/GoogleCloudStorage.feature:162 Nov 14, 2022 3:05:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 3:05:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a2365286-9d96-4d41-a0db-e02c98dce914 for {dataset_id} Nov 14, 2022 3:05:52 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-PipeFieldSeparator config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:05:52 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with Common IO CSV PipeFieldSeparator -> test # features/connectors/Google/GoogleCloudStorage.feature:169 Nov 14, 2022 3:06:08 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:06:08 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a2365286-9d96-4d41-a0db-e02c98dce914 for {dataset_id} Nov 14, 2022 3:06:08 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-PipeFieldSeparator config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:06:08 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with Common IO CSV SemicolonFieldSeparator # features/connectors/Google/GoogleCloudStorage.feature:176 Nov 14, 2022 3:06:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 3:06:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 27bfbc60-6dbb-4d98-a5a0-b84dd727ed5c for {dataset_id} Nov 14, 2022 3:06:30 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-SemicolonFieldSeparator config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:06:30 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with Common IO CSV SemicolonFieldSeparator -> test # features/connectors/Google/GoogleCloudStorage.feature:183 Nov 14, 2022 3:06:46 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:06:46 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 27bfbc60-6dbb-4d98-a5a0-b84dd727ed5c for {dataset_id} Nov 14, 2022 3:06:46 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-SemicolonFieldSeparator config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:06:46 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with Common IO CSV Advanced EscapeCharacter # features/connectors/Google/GoogleCloudStorage.feature:190 Nov 14, 2022 3:07:08 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 3:07:08 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 6fe0e3b5-ec4a-4b37-9b16-edd961604eff for {dataset_id} Nov 14, 2022 3:07:08 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-Advanced-EscapeCharacter config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:07:08 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with Common IO CSV Advanced EscapeCharacter -> test # features/connectors/Google/GoogleCloudStorage.feature:197 Nov 14, 2022 3:07:24 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:07:24 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 6fe0e3b5-ec4a-4b37-9b16-edd961604eff for {dataset_id} Nov 14, 2022 3:07:24 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-Advanced-EscapeCharacter config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:07:24 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: test -> Google Cloud Storage with Common IO CSV Advanced QuoteCharacter # features/connectors/Google/GoogleCloudStorage.feature:204 Nov 14, 2022 3:07:46 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 8e83bf33-7c4d-4ab0-8f3d-5cabb92aca91 for {os-test_id} Nov 14, 2022 3:07:46 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 95e4e918-791c-43d9-a6d7-f43848967a7c for {dataset_id} Nov 14, 2022 3:07:46 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-Advanced-QuoteCharacter config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:07:46 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GoogleCloudStorage Scenario: Google Cloud Storage with Common IO CSV Advanced QuoteCharacter -> test # features/connectors/Google/GoogleCloudStorage.feature:211 Nov 14, 2022 3:08:01 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:08:01 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 95e4e918-791c-43d9-a6d7-f43848967a7c for {dataset_id} Nov 14, 2022 3:08:02 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GoogleCloudStorage is : 201 When user adds pipeline using GoogleCloudStorage with dataset CommonIO-CSV-Advanced-QuoteCharacter config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:08:02 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test GooglePubSub components @component @smoke @GooglePubSub Scenario: test data -> PubSub with CSV # features/connectors/Google/GooglePubSub.feature:6 Nov 14, 2022 3:08:23 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 4670c342-cef4-4596-8831-179fc56605f3 for {os-test_id} Nov 14, 2022 3:08:23 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: e082742b-4001-4213-a67e-91e40561c877 for {dataset_id} Nov 14, 2022 3:08:24 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GooglePubSub is : 201 When user adds pipeline using GooglePubSub with dataset Csv config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:08:24 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GooglePubSub Scenario: test data -> PubSub with text # features/connectors/Google/GooglePubSub.feature:13 Nov 14, 2022 3:08:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 95a4826e-16db-45ca-a7dc-89acfbab7dc7 for {os-test_id} Nov 14, 2022 3:08:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a023013a-699c-4ece-b9c0-a35c389579c6 for {dataset_id} Nov 14, 2022 3:08:52 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GooglePubSub is : 201 When user adds pipeline using GooglePubSub with dataset Text config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:08:52 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GooglePubSub Scenario: test data -> PubSub with Avro # features/connectors/Google/GooglePubSub.feature:20 Nov 14, 2022 3:09:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 95a4826e-16db-45ca-a7dc-89acfbab7dc7 for {os-test_id} Nov 14, 2022 3:09:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cc8eaf8d-2329-43fe-a339-fa89791da1bf for {dataset_id} Nov 14, 2022 3:09:14 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GooglePubSub is : 201 When user adds pipeline using GooglePubSub with dataset Avro config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:09:14 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GooglePubSub Scenario: test data -> PubSub with JSON # features/connectors/Google/GooglePubSub.feature:27 Nov 14, 2022 3:09:36 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 95a4826e-16db-45ca-a7dc-89acfbab7dc7 for {os-test_id} Nov 14, 2022 3:09:36 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 3bed2058-6787-422c-a1c6-dcde7521482f for {dataset_id} Nov 14, 2022 3:09:36 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GooglePubSub is : 201 When user adds pipeline using GooglePubSub with dataset JSON config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:09:36 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @GooglePubSub Scenario: PubSub with CSV -> test # features/connectors/Google/GooglePubSub.feature:34 Nov 14, 2022 3:09:58 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:09:58 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: e082742b-4001-4213-a67e-91e40561c877 for {dataset_id} Nov 14, 2022 3:09:58 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GooglePubSub is : 201 When user adds pipeline using GooglePubSub with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:09:58 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 3:11:35 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component @GooglePubSub Scenario: PubSub with text -> test # features/connectors/Google/GooglePubSub.feature:42 Nov 14, 2022 3:11:38 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:11:38 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a023013a-699c-4ece-b9c0-a35c389579c6 for {dataset_id} Nov 14, 2022 3:11:38 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GooglePubSub is : 201 When user adds pipeline using GooglePubSub with dataset Text config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:11:38 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 3:13:11 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component @smoke @GooglePubSub Scenario: PubSub with Avro -> test # features/connectors/Google/GooglePubSub.feature:50 Nov 14, 2022 3:13:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:13:14 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cc8eaf8d-2329-43fe-a339-fa89791da1bf for {dataset_id} Nov 14, 2022 3:13:15 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GooglePubSub is : 201 When user adds pipeline using GooglePubSub with dataset Avro config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:13:15 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 3:14:48 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component @GooglePubSub Scenario: PubSub with JSON -> test # features/connectors/Google/GooglePubSub.feature:59 Nov 14, 2022 3:14:51 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:14:51 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 3bed2058-6787-422c-a1c6-dcde7521482f for {dataset_id} Nov 14, 2022 3:14:51 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using GooglePubSub is : 201 When user adds pipeline using GooglePubSub with dataset JSON config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:14:51 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 3:16:25 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component Feature: test HDFS components @component @HDFS Scenario: test data -> HDFS # features/connectors/HDFS.feature:5 Nov 14, 2022 3:16:28 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 3:16:28 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:16:28 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: d92ab37f-3408-41ab-b076-1f44b58a38e8 for {dataset_id} Nov 14, 2022 3:16:28 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using HDFS is : 201 When user adds pipeline using HDFS with dataset Csv config for output component with OverwriteMerge config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:16:28 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() java.lang.AssertionError: The pipeline does not run successfully, the last run status is [running, ] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkPipelineStatus(TPDAPIOperation.java:431) at steps.AddPipelineStep.checkPipelineStatus(AddPipelineStep.java:83) at ✽.the lastExecutionStatus for the pipeline is "finished"(features/connectors/HDFS.feature:8) @component @HDFS Scenario: HDFS -> test # features/connectors/HDFS.feature:11 Nov 14, 2022 3:20:32 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:20:32 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: d92ab37f-3408-41ab-b076-1f44b58a38e8 for {dataset_id} Nov 14, 2022 3:20:32 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using HDFS is : 201 When user adds pipeline using HDFS with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:20:33 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() java.lang.AssertionError: The pipeline does not run successfully, the last run status is [running, ] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkPipelineStatus(TPDAPIOperation.java:431) at steps.AddPipelineStep.checkPipelineStatus(AddPipelineStep.java:83) at ✽.the lastExecutionStatus for the pipeline is "finished"(features/connectors/HDFS.feature:14) @component Feature: test kafka components @component @Kafka Scenario: test data -> kafka with csv # features/connectors/Kafka.feature:5 Nov 14, 2022 3:24:37 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 3:24:37 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:24:37 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 612487c7-b636-43ef-9e44-b7119271c17d for {dataset_id} Nov 14, 2022 3:24:37 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Kafka is : 201 When user adds pipeline using Kafka with dataset Csv config for output component with RoundRobin config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:24:37 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @Kafka Scenario: test data -> kafka with json # features/connectors/Kafka.feature:11 Nov 14, 2022 3:24:52 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 3:24:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:24:52 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 471b8a81-a59b-4e40-a711-58e5aa38dea8 for {dataset_id} Nov 14, 2022 3:24:52 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Kafka is : 201 When user adds pipeline using Kafka with dataset Json config for output component with RoundRobin config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:24:52 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Kafka Scenario: test data -> kafka with avro # features/connectors/Kafka.feature:17 Nov 14, 2022 3:25:08 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 3:25:08 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:25:08 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 3559b9da-731d-4567-bc36-33a3211fd4d9 for {dataset_id} Nov 14, 2022 3:25:08 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Kafka is : 201 When user adds pipeline using Kafka with dataset Avro config for output component with RoundRobin config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:25:08 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @Kafka Scenario: kafka with csv -> test # features/connectors/Kafka.feature:23 Nov 14, 2022 3:25:23 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:25:23 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 612487c7-b636-43ef-9e44-b7119271c17d for {dataset_id} Nov 14, 2022 3:25:24 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Kafka is : 201 When user adds pipeline using Kafka with dataset Csv config for input component with Latest config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:25:24 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 3:26:57 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component Feature: test Kudu components @component @smoke @Kudu Scenario: test -> Kudu # features/connectors/Kudu.feature:5 Nov 14, 2022 3:27:00 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 3:27:00 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:27:00 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 09d0e873-2a2c-46b1-812b-3b220661ae0e for {dataset_id} Nov 14, 2022 3:27:00 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Kudu is : 201 When user adds pipeline using Kudu without dataset config for output component without config # AddPipelineStep.addPipeline(String,String) Nov 14, 2022 3:27:01 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @Kudu Scenario: Kudu -> test # features/connectors/Kudu.feature:11 Nov 14, 2022 3:27:16 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:27:16 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 09d0e873-2a2c-46b1-812b-3b220661ae0e for {dataset_id} Nov 14, 2022 3:27:16 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Kudu is : 201 When user adds pipeline using Kudu without dataset config for input component without config # AddPipelineStep.addPipeline(String,String) Nov 14, 2022 3:27:16 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test Local Connection components @component @smoke @LocalConnection Scenario: test -> Local Connection with CSV # features/connectors/LocalConnection.feature:5 Nov 14, 2022 3:27:31 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 3:27:31 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:27:31 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: deb9d4f0-3334-4f31-82c9-f3c24717311f for {dataset_id} Nov 14, 2022 3:27:32 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using LocalConnection is : 201 When user adds pipeline using LocalConnection with dataset Csv config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:27:32 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @LocalConnection Scenario: Local Connection with CSV -> test # features/connectors/LocalConnection.feature:11 Nov 14, 2022 3:27:53 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:27:53 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: deb9d4f0-3334-4f31-82c9-f3c24717311f for {dataset_id} Nov 14, 2022 3:27:53 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using LocalConnection is : 201 When user adds pipeline using LocalConnection with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:27:53 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test Marketo components @component @smoke @Marketo Scenario: marketo -> test # features/connectors/Marketo.feature:5 Nov 14, 2022 3:28:15 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:28:15 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 1b63c435-6f98-4b63-942b-57522ab905f7 for {dataset_id} Nov 14, 2022 3:28:15 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Marketo is : 201 When user adds pipeline using Marketo with dataset People config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:28:15 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @Marketo Scenario: test -> marketo # features/connectors/Marketo.feature:11 Nov 14, 2022 3:28:30 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 3:28:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:28:30 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 1b63c435-6f98-4b63-942b-57522ab905f7 for {dataset_id} Nov 14, 2022 3:28:31 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Marketo is : 201 When user adds pipeline using Marketo with dataset People config for output component with CreateOrUpdate config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:28:31 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test MongoDB components @component @smoke @MongoDB Scenario: MongoDB -> test # features/connectors/MongoDB.feature:5 Nov 14, 2022 3:28:46 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:28:46 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: d92547c3-9051-4194-80d1-4ee65b6ee6eb for {dataset_id} Nov 14, 2022 3:28:46 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MongoDB is : 201 When user adds pipeline using MongoDB with dataset Collection-Json config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:28:46 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @MongoDB Scenario: test -> MongoDB # features/connectors/MongoDB.feature:11 Nov 14, 2022 3:29:01 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 3:29:02 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:29:02 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: d92547c3-9051-4194-80d1-4ee65b6ee6eb for {dataset_id} Nov 14, 2022 3:29:02 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using MongoDB is : 201 When user adds pipeline using MongoDB with dataset Collection-Json config for output component with Insert config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:29:02 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke Feature: test NetSuite components @component @smoke @NetSuite Scenario: netsuite -> test # features/connectors/NetSuite.feature:5 Nov 14, 2022 3:29:17 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:29:17 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 423921e0-dd09-46fe-8af2-6fc39b8b2f3c for {dataset_id} Nov 14, 2022 3:29:17 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using NetSuite is : 201 When user adds pipeline using NetSuite without dataset config for input component without config # AddPipelineStep.addPipeline(String,String) Nov 14, 2022 3:29:17 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @NetSuite Scenario: test -> netsuite using update # features/connectors/NetSuite.feature:11 Nov 14, 2022 3:29:51 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 73eb00fc-501f-4bf2-8118-fde1b62bb235 for {os-test_id} Nov 14, 2022 3:29:51 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 423921e0-dd09-46fe-8af2-6fc39b8b2f3c for {dataset_id} Nov 14, 2022 3:29:51 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using NetSuite is : 201 When user adds pipeline using NetSuite without dataset config for output component with Update config # AddPipelineStep.addPipelineWithSetting(String,String,String) Nov 14, 2022 3:29:51 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test Pulsar components @component @smoke @Pulsar Scenario: test data -> Pulsar with CSV # features/connectors/Pulsar.feature:5 Nov 14, 2022 3:30:25 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 3:30:25 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:30:25 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cf17dd29-252e-4f01-9b3e-ac01cfae8256 for {dataset_id} Nov 14, 2022 3:30:25 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Pulsar is : 201 When user adds pipeline using Pulsar with dataset Csv config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:30:25 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @Pulsar Scenario: Pulsar with CSV -> test # features/connectors/Pulsar.feature:11 Nov 14, 2022 3:30:40 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:30:40 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cf17dd29-252e-4f01-9b3e-ac01cfae8256 for {dataset_id} Nov 14, 2022 3:30:41 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Pulsar is : 201 When user adds pipeline using Pulsar with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:30:41 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 3:32:14 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component Feature: test REST components @component @smoke @REST Scenario: test data -> rest # features/connectors/REST/REST.feature:5 Nov 14, 2022 3:32:23 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 3:32:23 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:32:23 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: c04eb6b3-f4b3-4636-926b-fb0ff2ebfa31 for {dataset_id} Nov 14, 2022 3:32:23 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using REST is : 201 When user adds pipeline using REST with dataset Batch config for output component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:32:24 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @REST Scenario: rest -> test # features/connectors/REST/REST.feature:11 Nov 14, 2022 3:32:39 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:32:39 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: c04eb6b3-f4b3-4636-926b-fb0ff2ebfa31 for {dataset_id} Nov 14, 2022 3:32:39 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using REST is : 201 When user adds pipeline using REST with dataset Batch config for input component with Batch config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:32:39 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @REST Scenario: rest using polling -> test # features/connectors/REST/REST.feature:17 Nov 14, 2022 3:32:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:32:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: b1a70742-2b8c-4bdd-b3eb-53e5dec7464e for {dataset_id} Nov 14, 2022 3:32:55 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using REST is : 201 When user adds pipeline using REST with dataset Streaming config for input component with Streaming config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:32:55 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the pipeline status is running still after waiting for 90 seconds # AddPipelineStep.runningStatus() Nov 14, 2022 3:34:28 PM actions.TPDAPIOperation killPipeline INFO: Status code for running pipeline is 200 When user kills the running pipeline # AddPipelineStep.killPipeline() @component @REST Scenario: rest with header parameter -> test # features/connectors/REST/REST.feature:25 Nov 14, 2022 3:34:31 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:34:31 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 3153928b-9a4e-40f3-97a0-fcd46b9479d3 for {dataset_id} Nov 14, 2022 3:34:31 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using REST is : 201 When user adds pipeline using REST with dataset Header-Parameter config for input component with Batch config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:34:31 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @REST Scenario: rest with URL parameter -> test # features/connectors/REST/REST.feature:32 Nov 14, 2022 3:34:47 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:34:47 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 283a8aa5-ac19-4377-bf95-827aa1a708fe for {dataset_id} Nov 14, 2022 3:34:47 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using REST is : 201 When user adds pipeline using REST with dataset URL-Parameter config for input component with Batch config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:34:47 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test REST_Authentication components @component @REST Scenario: REST with basic authentication -> test # features/connectors/REST/REST_Authentication_Basic.feature:6 Nov 14, 2022 3:35:02 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:35:02 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 84b07407-3cb8-4d77-b2b2-9ff92991ce39 for {dataset_id} Nov 14, 2022 3:35:02 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using REST_Authentication_Basic is : 201 When user adds pipeline using REST_Authentication_Basic with dataset Basic_Auth config for input component with Batch config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:35:02 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test REST_Authentication components @component @REST Scenario: REST with digest authentication -> test # features/connectors/REST/REST_Authentication_Digest.feature:6 Nov 14, 2022 3:35:18 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:35:18 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 175748d4-0504-46aa-a420-a2fee18e69c7 for {dataset_id} Nov 14, 2022 3:35:18 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using REST_Authentication_Digest is : 201 When user adds pipeline using REST_Authentication_Digest with dataset Digest_Auth config for input component with Batch config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:35:18 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test REST_HTTPBin components @component @REST Scenario: rest with path parameter -> test # features/connectors/REST/REST_HTTPBin.feature:6 Nov 14, 2022 3:35:33 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:35:33 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: a6150171-85c4-45b3-9f57-a2a2b477a982 for {dataset_id} Nov 14, 2022 3:35:33 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using REST_HTTPBin is : 201 When user adds pipeline using REST_HTTPBin with dataset Path-Parameter config for input component with Batch config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:35:34 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @REST Scenario: rest with body parameter -> test # features/connectors/REST/REST_HTTPBin.feature:13 Nov 14, 2022 3:35:49 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:35:49 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: f3342f18-b0bf-4fdf-80ff-7cab26209270 for {dataset_id} Nov 14, 2022 3:35:49 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using REST_HTTPBin is : 201 When user adds pipeline using REST_HTTPBin with dataset Body-Parameter config for input component with Batch config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:35:49 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test SFTP components @component @smoke @SFTP Scenario: test -> SFTP with CSV # features/connectors/SFTP.feature:5 Nov 14, 2022 3:36:04 PM utils.payloadAction.TPDPayloadAction updatePLDependency INFO: No dataset setting found in os-test_id.properties. Default dataset - "Test" will be use for {os-test_id} Nov 14, 2022 3:36:04 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:36:04 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 5b296836-d728-4ddf-949a-19d1e6185e81 for {dataset_id} Nov 14, 2022 3:36:05 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using SFTP is : 201 When user adds pipeline using SFTP with dataset Csv config for output component with Size config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:36:05 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @SFTP Scenario: SFTP with CSV -> test # features/connectors/SFTP.feature:11 Nov 14, 2022 3:36:26 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:36:26 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 5b296836-d728-4ddf-949a-19d1e6185e81 for {dataset_id} Nov 14, 2022 3:36:26 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using SFTP is : 201 When user adds pipeline using SFTP with dataset Csv config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:36:26 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component Feature: test Salesforce components @component @smoke @Salesforce Scenario: salesforce -> test # features/connectors/Salesforce.feature:5 Nov 14, 2022 3:36:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:36:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 37bab68f-149c-449e-91cd-b7affc57225e for {dataset_id} Nov 14, 2022 3:36:55 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Salesforce is : 201 When user adds pipeline using Salesforce with dataset Module config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:36:55 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Salesforce Scenario: salesforce query -> test # features/connectors/Salesforce.feature:11 Nov 14, 2022 3:37:16 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:37:16 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 9ae7b439-06f4-4fff-ba92-f1ba53e4743a for {dataset_id} Nov 14, 2022 3:37:17 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Salesforce is : 201 When user adds pipeline using Salesforce with dataset Query config for input component with Query config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:37:17 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @smoke @Salesforce Scenario: test -> salesforce # features/connectors/Salesforce.feature:17 Nov 14, 2022 3:37:38 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: ad3868a4-5207-4892-a0bd-0ebe060c7546 for {os-test_id} Nov 14, 2022 3:37:38 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 37bab68f-149c-449e-91cd-b7affc57225e for {dataset_id} Nov 14, 2022 3:37:39 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Salesforce is : 201 When user adds pipeline using Salesforce with dataset Module config for output component with Insert config # AddPipelineStep.addPipelineWithSetting(String,String,String,String) Nov 14, 2022 3:37:39 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Workday Feature: test Workday components @component @Workday Scenario: workdayRaaS -> test # features/connectors/Workday.feature:10 Nov 14, 2022 3:37:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:37:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: 7fd47f1b-cd08-41f3-be47-671fbd5f0f44 for {dataset_id} Nov 14, 2022 3:37:55 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Workday is : 201 When user adds pipeline using Workday with dataset RaaS config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:37:55 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @Workday Scenario: workdayWSSA -> test # features/connectors/Workday.feature:15 Nov 14, 2022 3:38:16 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:38:16 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: b40b0939-3608-4c01-bad9-094914e94d9f for {dataset_id} Nov 14, 2022 3:38:17 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using Workday is : 201 When user adds pipeline using Workday with dataset WSSA config for input component without config # AddPipelineStep.addPipeline(String,String,String) Nov 14, 2022 3:38:17 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component Feature: all processors @processor @component @smoke Scenario: test dataset -> Filter -> test # features/processors/Processors.feature:5 Nov 14, 2022 3:38:38 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:38:38 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-filter is : 201 When user adds pipeline using proc-filter processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:38:39 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @smoke Scenario: test dataset -> Field Selector -> test # features/processors/Processors.feature:11 Nov 14, 2022 3:38:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:38:54 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-fieldselector is : 201 When user adds pipeline using proc-fieldselector processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:38:54 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component Scenario: test dataset -> Normalize -> test # features/processors/Processors.feature:16 Nov 14, 2022 3:39:10 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:39:10 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-normalize is : 201 When user adds pipeline using proc-normalize processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:39:10 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @smoke Scenario: test dataset -> Python -> S3 # features/processors/Processors.feature:22 Nov 14, 2022 3:39:26 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:39:26 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-normalize is : 201 When user adds pipeline using proc-normalize processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:39:26 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component Scenario: test dataset -> Replica -> test # features/processors/Processors.feature:27 Nov 14, 2022 3:39:41 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:39:42 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-replica is : 201 When user adds pipeline using proc-replica processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:39:42 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @join Scenario: test dataset -> Join -> test # features/processors/Processors.feature:33 Nov 14, 2022 3:39:57 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:39:57 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-join is : 201 When user adds pipeline using proc-join processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:39:58 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @smoke Scenario: test dataset -> Type Converter -> test # features/processors/Processors.feature:39 Nov 14, 2022 3:40:19 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:40:19 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-typeconverter is : 201 When user adds pipeline using proc-typeconverter processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:40:20 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @smoke Scenario: test dataset -> Aggregate -> test # features/processors/Processors.feature:45 Nov 14, 2022 3:40:35 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:40:35 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-aggregate is : 201 When user adds pipeline using proc-aggregate processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:40:35 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @smoke Scenario: test dataset -> Split -> test # features/processors/Processors.feature:51 Nov 14, 2022 3:40:51 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:40:51 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-split is : 201 When user adds pipeline using proc-split processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:40:51 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @2020-07 Scenario: test dataset -> Math -> test # features/processors/Processors.feature:57 Nov 14, 2022 3:41:07 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:41:07 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-math is : 201 When user adds pipeline using proc-math processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:41:07 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component Scenario: test dataset -> Phones -> test # features/processors/Processors.feature:62 Nov 14, 2022 3:41:23 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:41:23 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-phones is : 201 When user adds pipeline using proc-phones processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:41:23 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @smoke Scenario: test dataset -> DataCleansing -> test # features/processors/Processors.feature:68 Nov 14, 2022 3:41:38 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:41:39 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-datacleansing is : 201 When user adds pipeline using proc-datacleansing processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:41:39 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @smoke Scenario: test dataset -> Strings -> test # features/processors/Processors.feature:74 Nov 14, 2022 3:41:54 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:41:54 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-strings is : 201 When user adds pipeline using proc-strings processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:41:55 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @release @smoke Scenario: test dataset -> Dates -> test # features/processors/Processors.feature:80 Nov 14, 2022 3:42:10 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:42:10 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-dates is : 201 When user adds pipeline using proc-dates processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:42:10 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @2020-07-EAP @2020-07 Scenario: test dataset -> DataHashing -> test # features/processors/Processors.feature:86 Nov 14, 2022 3:42:32 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:42:32 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-datahashing is : 201 When user adds pipeline using proc-datahashing processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:42:32 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @2020-09-EAP Scenario: test dataset -> Numbers -> test # features/processors/Processors.feature:92 Nov 14, 2022 3:42:48 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:42:48 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-numbers is : 201 When user adds pipeline using proc-numbers processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:42:48 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @2020-09-EAP Scenario: test dataset -> Fields Concatenator -> test # features/processors/Processors.feature:98 Nov 14, 2022 3:43:04 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:43:04 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-fieldsconcatenator is : 201 When user adds pipeline using proc-fieldsconcatenator processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:43:04 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component Scenario: test dataset -> Data Sampling -> test # features/processors/Processors.feature:103 Nov 14, 2022 3:43:20 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:43:20 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-datasampling is : 201 When user adds pipeline using proc-datasampling processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:43:20 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @processor @component @RunPreparation Scenario: test dataset -> Run preparation -> test # features/processors/Processors.feature:109 When user adds a preparation # UtilStep.addPreparation() Nov 14, 2022 3:43:41 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:43:41 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-runpreparation is : 201 When user adds pipeline using proc-runpreparation processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:43:42 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is "finished" # AddPipelineStep.checkPipelineStatus() @component @processor @smoke Feature: Python2 security @component @processor @smoke Scenario: import library denies # features/processors/Python2.feature:4 Nov 14, 2022 3:43:57 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:43:57 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-python2-subprocess is : 201 When user adds pipeline using proc-python2-subprocess processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:43:58 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is Error with message UnknownReason # AddPipelineStep.checkFailedPipeline(String) @component @processor @smoke Scenario: reading file denies # features/processors/Python2.feature:9 Nov 14, 2022 3:44:19 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:44:19 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-python2-reading-file is : 201 When user adds pipeline using proc-python2-reading-file processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:44:20 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is Error with message UnknownReason # AddPipelineStep.checkFailedPipeline(String) @component @processor @smoke @skipOnat @skipOnqa @skipOnstaging @skipOnaz-staging @skipOnap @skipOnus @skipOneu @skipOnap @skipOnus-west Feature: Python3 security @component @processor @smoke @skipOnat @skipOnqa @skipOnstaging @skipOnaz-staging @skipOnap @skipOnus @skipOneu @skipOnap @skipOnus-west Scenario: deny telnetlib library # features/processors/Python3.feature:4 Nov 14, 2022 3:44:41 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:44:41 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-python3-telnetlib is : 201 When user adds pipeline using proc-python3-telnetlib processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:44:42 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is Error with message The import 'telnetlib' cannot be used # AddPipelineStep.checkFailedPipeline(String) java.lang.AssertionError: The error message is not the expected, the actual message is [finished, ] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkFailedPipeline(TPDAPIOperation.java:449) at steps.AddPipelineStep.checkFailedPipeline(AddPipelineStep.java:87) at ✽.the lastExecutionStatus for the pipeline is Error with message The import 'telnetlib' cannot be used(features/processors/Python3.feature:7) @component @processor @smoke @skipOnat @skipOnqa @skipOnstaging @skipOnaz-staging @skipOnap @skipOnus @skipOneu @skipOnap @skipOnus-west Scenario: deny socket library # features/processors/Python3.feature:9 Nov 14, 2022 3:46:45 PM utils.payloadAction.TPDPayloadAction replaceidentifier INFO: Using Dataset: cdcc2816-925d-43aa-b644-03c25b5e3395 for {os-test_id} Nov 14, 2022 3:46:45 PM actions.TPDAPIOperation userAddsPipeline INFO: Status code for adding pipeline using proc-python3-socket is : 201 When user adds pipeline using proc-python3-socket processor to test processor component without config # AddPipelineStep.addPipelineWithSetting(String) Nov 14, 2022 3:46:45 PM actions.TPDAPIOperation runPipeline INFO: Status code for running pipeline is 200 When run the inject pipeline # AddPipelineStep.runPipeline() Then the lastExecutionStatus for the pipeline is Error with message The import 'socket' cannot be used # AddPipelineStep.checkFailedPipeline(String) java.lang.AssertionError: The error message is not the expected, the actual message is [finished, ] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkFailedPipeline(TPDAPIOperation.java:449) at steps.AddPipelineStep.checkFailedPipeline(AddPipelineStep.java:87) at ✽.the lastExecutionStatus for the pipeline is Error with message The import 'socket' cannot be used(features/processors/Python3.feature:12) @sample @component Feature: Amazon service sample test @sample @component Scenario: check Kinesis sample # features/sample/AmazonService-Sample.feature:4 When user refresh AmazonKinesis without dataset config sample # SampleStep.userRefreshSample(String) Nov 14, 2022 3:49:00 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset AmazonKinesis without dataset config sample successfully # SampleStep.getSample(String) @sample @component @Amazon-S3 @smoke Scenario: check s3 with avro sample # features/sample/AmazonService-Sample.feature:9 When user refresh AmazonS3 with Avro dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:49:12 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset AmazonS3 with Avro dataset config sample successfully # SampleStep.getSample(String,String) @sample @component @Amazon-S3 Scenario: check s3 with excel sample # features/sample/AmazonService-Sample.feature:14 When user refresh AmazonS3 with Excel dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:49:24 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset AmazonS3 with Excel dataset config sample successfully # SampleStep.getSample(String,String) @sample @component @Amazon-S3 Scenario: check s3 with parquet sample # features/sample/AmazonService-Sample.feature:19 When user refresh AmazonS3 with Parquet dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:49:36 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset AmazonS3 with Parquet dataset config sample successfully # SampleStep.getSample(String,String) @sample @component Feature: Azure Service sample test @sample @component @smoke Scenario: check Azure Event Hubs sample # features/sample/AzureService-Sample.feature:12 When user refresh AzureEventHubs with Csv dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:49:47 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:49:56 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset AzureEventHubs with Csv dataset config sample successfully # SampleStep.getSample(String,String) @sample @component @smoke Scenario: check azure blob with avro sample # features/sample/AzureService-Sample.feature:17 When user refresh AzureBlobStorage with Avro dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:50:07 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset AzureBlobStorage with Avro dataset config sample successfully # SampleStep.getSample(String,String) @sample @component Scenario: check azure blob with parquet sample # features/sample/AzureService-Sample.feature:21 When user refresh AzureBlobStorage with Parquet dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:50:19 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset AzureBlobStorage with Parquet dataset config sample successfully # SampleStep.getSample(String,String) @sample @component @smoke Scenario: check azure datalake with avro sample # features/sample/AzureService-Sample.feature:26 When user refresh AzureDataLakeStorageGen2 with Avro dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:50:31 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset AzureDataLakeStorageGen2 with Avro dataset config sample successfully # SampleStep.getSample(String,String) @sample @component Scenario: check azure datalake with json sample # features/sample/AzureService-Sample.feature:30 When user refresh AzureDataLakeStorageGen2 with Json dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:50:42 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset AzureDataLakeStorageGen2 with Json dataset config sample successfully # SampleStep.getSample(String,String) @sample @component Scenario: check azure datalake with parquet sample # features/sample/AzureService-Sample.feature:34 When user refresh AzureDataLakeStorageGen2 with Parquet dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:50:54 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset AzureDataLakeStorageGen2 with Parquet dataset config sample successfully # SampleStep.getSample(String,String) @sample @component Scenario: check azure datalake with csv sample # features/sample/AzureService-Sample.feature:38 When user refresh AzureDataLakeStorageGen2 with Csv dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:51:06 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset AzureDataLakeStorageGen2 with Csv dataset config sample successfully # SampleStep.getSample(String,String) @sample @component Feature: Common Service sample test @sample @component @smoke Scenario: check marketo sample # features/sample/CommonService-Sample.feature:5 When user refresh Marketo with People dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:51:17 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset Marketo with People dataset config sample successfully # SampleStep.getSample(String,String) @sample @component @smoke Scenario: check salesforce sample # features/sample/CommonService-Sample.feature:10 When user refresh Salesforce with Module dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:51:29 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset Salesforce with Module dataset config sample successfully # SampleStep.getSample(String,String) @sample @component @smoke Scenario: check netsuite sample # features/sample/CommonService-Sample.feature:15 When user refresh NetSuite without dataset config sample # SampleStep.userRefreshSample(String) Nov 14, 2022 3:51:41 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:51:49 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset NetSuite without dataset config sample successfully # SampleStep.getSample(String) @sample @component Scenario: check HDFS sample # features/sample/CommonService-Sample.feature:27 When user refresh HDFS with Csv dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:52:01 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:52:09 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:52:17 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:52:25 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:52:34 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:52:42 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:52:53 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:53:02 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:53:10 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:53:18 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:53:26 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:53:34 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:53:46 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:53:54 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:54:02 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:54:11 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:54:19 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:54:27 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:54:27 PM actions.TPDAPIOperation checkSampleResult SEVERE: Error Messsage: {"timestamp":1668441262773,"status":404,"error":"Not found","message":"Failed to fetch the sample from dataset d92ab37f-3408-41ab-b076-1f44b58a38e8."} Then user gets dataset HDFS with Csv dataset config sample successfully # SampleStep.getSample(String,String) java.lang.AssertionError: Status Check expected:<200> but was:<404> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:645) at utils.TPDAPIRequest.checkStatus(TPDAPIRequest.java:117) at actions.TPDAPIOperation.checkSampleResult(TPDAPIOperation.java:346) at steps.SampleStep.getSample(SampleStep.java:31) at ✽.user gets dataset HDFS with Csv dataset config sample successfully(features/sample/CommonService-Sample.feature:29) @sample @component @smoke Scenario: check FTP sample # features/sample/CommonService-Sample.feature:32 When user refresh FTP with Csv dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:54:39 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset FTP with Csv dataset config sample successfully # SampleStep.getSample(String,String) @sample @component @smoke Scenario: check SFTP sample # features/sample/CommonService-Sample.feature:37 When user refresh SFTP with Csv dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:54:51 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:54:59 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset SFTP with Csv dataset config sample successfully # SampleStep.getSample(String,String) @sample @component @smoke Scenario: check elasticsearch sample # features/sample/CommonService-Sample.feature:42 When user refresh ElasticSearch without dataset config sample # SampleStep.userRefreshSample(String) Nov 14, 2022 3:55:11 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset ElasticSearch without dataset config sample successfully # SampleStep.getSample(String) @sample @component @smoke Scenario: check rest sample # features/sample/CommonService-Sample.feature:47 When user refresh REST with Batch dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:55:23 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset REST with Batch dataset config sample successfully # SampleStep.getSample(String,String) @sample @component @smoke Scenario: check kafka sample # features/sample/CommonService-Sample.feature:52 When user refresh Kafka with Csv dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:55:34 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:55:43 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:55:51 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:55:59 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:56:07 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:56:15 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset Kafka with Csv dataset config sample successfully # SampleStep.getSample(String,String) @sample @component @smoke Scenario: check REST with polling sample # features/sample/CommonService-Sample.feature:57 When user refresh REST with Streaming dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:56:27 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:56:35 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:56:43 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:56:51 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:57:00 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:57:08 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:57:19 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:57:28 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:57:36 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:57:44 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:57:52 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:58:00 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:58:12 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:58:20 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:58:28 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:58:37 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:58:45 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:58:53 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:58:53 PM actions.TPDAPIOperation checkSampleResult SEVERE: Error Messsage: {"timestamp":1668441528358,"status":404,"error":"Not found","message":"Failed to fetch the sample from dataset b1a70742-2b8c-4bdd-b3eb-53e5dec7464e."} Then user gets dataset REST with Streaming dataset config sample successfully # SampleStep.getSample(String,String) java.lang.AssertionError: Status Check expected:<200> but was:<404> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:645) at utils.TPDAPIRequest.checkStatus(TPDAPIRequest.java:117) at actions.TPDAPIOperation.checkSampleResult(TPDAPIOperation.java:346) at steps.SampleStep.getSample(SampleStep.java:31) at ✽.user gets dataset REST with Streaming dataset config sample successfully(features/sample/CommonService-Sample.feature:59) @sample @component @smoke Scenario: check MongoDB sample # features/sample/CommonService-Sample.feature:62 When user refresh MongoDB with Collection-Json dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:59:05 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset MongoDB with Collection-Json dataset config sample successfully # SampleStep.getSample(String,String) @sample @component @LocalConnection @smoke Scenario: check Local Connection sample # features/sample/CommonService-Sample.feature:67 When user refresh LocalConnection with Csv dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:59:17 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset LocalConnection with Csv dataset config sample successfully # SampleStep.getSample(String,String) @sample @smoke @component @Database Feature: Database sample test @sample @smoke @component @Database Scenario: check amazon aurora sample # features/sample/Database-Sample.feature:4 When user refresh Aurora with Table dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:59:29 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset Aurora with Table dataset config sample successfully # SampleStep.getSample(String,String) @sample @smoke @component @Database Scenario: check azure dwh sql sample # features/sample/Database-Sample.feature:8 When user refresh AzureSynapse with Table dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 3:59:40 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 3:59:49 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset AzureSynapse with Table dataset config sample successfully # SampleStep.getSample(String,String) @sample @smoke @component @Database Scenario: check amazon mariadb sample # features/sample/Database-Sample.feature:12 When user refresh MariaDB with Table dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 4:00:00 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset MariaDB with Table dataset config sample successfully # SampleStep.getSample(String,String) @sample @smoke @component @Database Scenario: check amazon mssql sample # features/sample/Database-Sample.feature:16 When user refresh MSSQL with Table dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 4:00:13 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset MSSQL with Table dataset config sample successfully # SampleStep.getSample(String,String) @sample @smoke @component @Database Scenario: check amazon mysql sample # features/sample/Database-Sample.feature:20 When user refresh MySQL with Table dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 4:00:24 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset MySQL with Table dataset config sample successfully # SampleStep.getSample(String,String) @sample @smoke @component @Database Scenario: check amazon postgresql sample # features/sample/Database-Sample.feature:24 When user refresh PostgreSQL with Table dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 4:00:36 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset PostgreSQL with Table dataset config sample successfully # SampleStep.getSample(String,String) @sample @smoke @component @Database Scenario: check amazon redshift sample # features/sample/Database-Sample.feature:28 When user refresh Redshift with Table dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 4:00:48 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset Redshift with Table dataset config sample successfully # SampleStep.getSample(String,String) @sample @smoke @component @Database Scenario: check SnowFlake sample # features/sample/Database-Sample.feature:32 When user refresh Snowflake with Table dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 4:00:59 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 4:01:08 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset Snowflake with Table dataset config sample successfully # SampleStep.getSample(String,String) @sample @smoke @component Feature: Google Service sample test @sample @smoke @component Scenario: check GCS sample # features/sample/GoogleService-Sample.feature:4 When user refresh GoogleCloudStorage with Csv dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 4:01:19 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset GoogleCloudStorage with Csv dataset config sample successfully # SampleStep.getSample(String,String) @sample @smoke @component Scenario: check pubsub sample # features/sample/GoogleService-Sample.feature:8 When user refresh GooglePubSub with Csv dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 4:01:31 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Nov 14, 2022 4:01:39 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset GooglePubSub with Csv dataset config sample successfully # SampleStep.getSample(String,String) @sample @smoke @component Scenario: check BigQuery sample # features/sample/GoogleService-Sample.feature:12 When user refresh GoogleBigQuery with Table dataset config sample # SampleStep.userRefreshSample(String,String) Nov 14, 2022 4:01:51 PM actions.TPDAPIOperation waitForSample SEVERE: sleep 5s Then user gets dataset GoogleBigQuery with Table dataset config sample successfully # SampleStep.getSample(String,String) Failed scenarios: features/connectors/Database/DeltaLake.feature:5 # test data -> DeltaLake features/connectors/HDFS.feature:5 # test data -> HDFS features/connectors/HDFS.feature:11 # HDFS -> test features/processors/Python3.feature:4 # deny telnetlib library features/processors/Python3.feature:9 # deny socket library features/sample/CommonService-Sample.feature:27 # check HDFS sample features/sample/CommonService-Sample.feature:57 # check REST with polling sample 251 Scenarios (7 failed, 244 passed) 733 Steps (7 failed, 726 passed) 126m6.117s java.lang.AssertionError: The pipeline does not run successfully, the last run status is [error, (com.simba.spark.support.exceptions.ErrorException) [Simba][SparkJDBCDriver](700100) Connection timeout expired. Details: None.] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkPipelineStatus(TPDAPIOperation.java:431) at steps.AddPipelineStep.checkPipelineStatus(AddPipelineStep.java:83) at ✽.the lastExecutionStatus for the pipeline is "finished"(features/connectors/Database/DeltaLake.feature:8) java.lang.AssertionError: The pipeline does not run successfully, the last run status is [running, ] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkPipelineStatus(TPDAPIOperation.java:431) at steps.AddPipelineStep.checkPipelineStatus(AddPipelineStep.java:83) at ✽.the lastExecutionStatus for the pipeline is "finished"(features/connectors/HDFS.feature:8) java.lang.AssertionError: The pipeline does not run successfully, the last run status is [running, ] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkPipelineStatus(TPDAPIOperation.java:431) at steps.AddPipelineStep.checkPipelineStatus(AddPipelineStep.java:83) at ✽.the lastExecutionStatus for the pipeline is "finished"(features/connectors/HDFS.feature:14) java.lang.AssertionError: The error message is not the expected, the actual message is [finished, ] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkFailedPipeline(TPDAPIOperation.java:449) at steps.AddPipelineStep.checkFailedPipeline(AddPipelineStep.java:87) at ✽.the lastExecutionStatus for the pipeline is Error with message The import 'telnetlib' cannot be used(features/processors/Python3.feature:7) java.lang.AssertionError: The error message is not the expected, the actual message is [finished, ] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkFailedPipeline(TPDAPIOperation.java:449) at steps.AddPipelineStep.checkFailedPipeline(AddPipelineStep.java:87) at ✽.the lastExecutionStatus for the pipeline is Error with message The import 'socket' cannot be used(features/processors/Python3.feature:12) java.lang.AssertionError: Status Check expected:<200> but was:<404> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:645) at utils.TPDAPIRequest.checkStatus(TPDAPIRequest.java:117) at actions.TPDAPIOperation.checkSampleResult(TPDAPIOperation.java:346) at steps.SampleStep.getSample(SampleStep.java:31) at ✽.user gets dataset HDFS with Csv dataset config sample successfully(features/sample/CommonService-Sample.feature:29) java.lang.AssertionError: Status Check expected:<200> but was:<404> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:645) at utils.TPDAPIRequest.checkStatus(TPDAPIRequest.java:117) at actions.TPDAPIOperation.checkSampleResult(TPDAPIOperation.java:346) at steps.SampleStep.getSample(SampleStep.java:31) at ✽.user gets dataset REST with Streaming dataset config sample successfully(features/sample/CommonService-Sample.feature:59) Tests run: 251, Failures: 7, Errors: 0, Skipped: 0, Time elapsed: 7,567.128 sec <<< FAILURE! test data -> DeltaLake(test DeltaLake components) Time elapsed: 184.212 sec <<< FAILURE! java.lang.AssertionError: The pipeline does not run successfully, the last run status is [error, (com.simba.spark.support.exceptions.ErrorException) [Simba][SparkJDBCDriver](700100) Connection timeout expired. Details: None.] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkPipelineStatus(TPDAPIOperation.java:431) at steps.AddPipelineStep.checkPipelineStatus(AddPipelineStep.java:83) at ✽.the lastExecutionStatus for the pipeline is "finished"(features/connectors/Database/DeltaLake.feature:8) test data -> HDFS(test HDFS components) Time elapsed: 244.355 sec <<< FAILURE! java.lang.AssertionError: The pipeline does not run successfully, the last run status is [running, ] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkPipelineStatus(TPDAPIOperation.java:431) at steps.AddPipelineStep.checkPipelineStatus(AddPipelineStep.java:83) at ✽.the lastExecutionStatus for the pipeline is "finished"(features/connectors/HDFS.feature:8) HDFS -> test(test HDFS components) Time elapsed: 244.553 sec <<< FAILURE! java.lang.AssertionError: The pipeline does not run successfully, the last run status is [running, ] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkPipelineStatus(TPDAPIOperation.java:431) at steps.AddPipelineStep.checkPipelineStatus(AddPipelineStep.java:83) at ✽.the lastExecutionStatus for the pipeline is "finished"(features/connectors/HDFS.feature:14) deny telnetlib library(Python3 security) Time elapsed: 123.43 sec <<< FAILURE! java.lang.AssertionError: The error message is not the expected, the actual message is [finished, ] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkFailedPipeline(TPDAPIOperation.java:449) at steps.AddPipelineStep.checkFailedPipeline(AddPipelineStep.java:87) at ✽.the lastExecutionStatus for the pipeline is Error with message The import 'telnetlib' cannot be used(features/processors/Python3.feature:7) deny socket library(Python3 security) Time elapsed: 123.519 sec <<< FAILURE! java.lang.AssertionError: The error message is not the expected, the actual message is [finished, ] at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at actions.TPDAPIOperation.checkFailedPipeline(TPDAPIOperation.java:449) at steps.AddPipelineStep.checkFailedPipeline(AddPipelineStep.java:87) at ✽.the lastExecutionStatus for the pipeline is Error with message The import 'socket' cannot be used(features/processors/Python3.feature:12) check HDFS sample(Common Service sample test) Time elapsed: 158.064 sec <<< FAILURE! java.lang.AssertionError: Status Check expected:<200> but was:<404> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:645) at utils.TPDAPIRequest.checkStatus(TPDAPIRequest.java:117) at actions.TPDAPIOperation.checkSampleResult(TPDAPIOperation.java:346) at steps.SampleStep.getSample(SampleStep.java:31) at ✽.user gets dataset HDFS with Csv dataset config sample successfully(features/sample/CommonService-Sample.feature:29) check REST with polling sample(Common Service sample test) Time elapsed: 157.62 sec <<< FAILURE! java.lang.AssertionError: Status Check expected:<200> but was:<404> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:645) at utils.TPDAPIRequest.checkStatus(TPDAPIRequest.java:117) at actions.TPDAPIOperation.checkSampleResult(TPDAPIOperation.java:346) at steps.SampleStep.getSample(SampleStep.java:31) at ✽.user gets dataset REST with Streaming dataset config sample successfully(features/sample/CommonService-Sample.feature:59) Results : Failed tests: test data -> DeltaLake(test DeltaLake components): The pipeline does not run successfully, the last run status is [error, (com.simba.spark.support.exceptions.ErrorException) [Simba][SparkJDBCDriver](700100) Connection timeout expired. Details: None.] test data -> HDFS(test HDFS components): The pipeline does not run successfully, the last run status is [running, ] HDFS -> test(test HDFS components): The pipeline does not run successfully, the last run status is [running, ] deny telnetlib library(Python3 security): The error message is not the expected, the actual message is [finished, ] deny socket library(Python3 security): The error message is not the expected, the actual message is [finished, ] check HDFS sample(Common Service sample test): Status Check expected:<200> but was:<404> check REST with polling sample(Common Service sample test): Status Check expected:<200> but was:<404> Tests run: 251, Failures: 7, Errors: 0, Skipped: 0 [ERROR] There are test failures. Please refer to /home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/target/surefire-reports for the individual test results. [INFO] ------------------------------------------------------------------------ [INFO] BUILD SUCCESS [INFO] ------------------------------------------------------------------------ [INFO] Total time: 02:06 h [INFO] Finished at: 2022-11-14T16:01:51Z [INFO] ------------------------------------------------------------------------ [Pipeline] } [Pipeline] // script [Pipeline] } [Pipeline] // stage [Pipeline] stage [Pipeline] { (Declarative: Post Actions) [Pipeline] junit Recording test results [Checks API] No suitable checks publisher found. [Pipeline] cucumber [CucumberReport] Using Cucumber Reports version 5.7.3 [CucumberReport] JSON report directory is "/home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/target/surefire-reports" [CucumberReport] Copied 0 properties files from workspace "/home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/target/surefire-reports" to reports directory "/var/jenkins_home/jobs/QA/jobs/component-api-test-new/builds/50/cucumber-html-reports/.cache" [CucumberReport] Copied 1 files from workspace "/home/jenkins/agent/workspace/QA/component-api-test-new/data-processing-runtime/data-processing-runtime-qa/dpruntime-tests-api-components/target/surefire-reports" to reports directory "/var/jenkins_home/jobs/QA/jobs/component-api-test-new/builds/50/cucumber-html-reports/.cache" [CucumberReport] Processing 1 json files: [CucumberReport] /var/jenkins_home/jobs/QA/jobs/component-api-test-new/builds/50/cucumber-html-reports/.cache/cucumberOriginal.json [CucumberReport] Found 7 failed steps, while expected at most 0 [CucumberReport] Build status is left unchanged [Pipeline] slackSend Slack Send Pipeline step running, values are - baseUrl: , teamDomain: talend, channel: dp-runtime-qa-ci-result, color: warning, botUser: true, tokenCredentialId: slack-token, notifyCommitters: false, iconEmoji: , username: , timestamp: [Pipeline] } [Pipeline] // stage [Pipeline] } [Pipeline] // timeout [Pipeline] } [Pipeline] // ansiColor [Pipeline] } [Pipeline] // withEnv [Pipeline] } [Pipeline] // container [Pipeline] } [Pipeline] // node [Pipeline] } [Pipeline] // podTemplate [Pipeline] End of Pipeline Finished: UNSTABLE