Data unavailable error message while running script from Runner - postman

I am trying to run my postman collection with the help of runner option while running the collection I am getting "Data unavailable" Error message and my script stop there itself.
Can anyone please guide me for same.

This is an ongoing issue from postman's end. For workaround you can delete your history and then run your collection.

Related

Google Cloud Run runs over and over again in error

I'm using Cloud Run that is getting triggered by a pubsub message.
But when this Cloud Run code gets an error it does re-run the application over and over again.
This seems unnecessary now when testing because I see the error in the log and doesn't need the code to re-run.
Where can I turn this off?
I'm using Node JS.
You can purge your PubSub push subscription, or delete it.
Solved it short term by enclosing the whole code block by a try/catch and then always be sure to a throw err to catch the error.
After that instead of returning a 400 status in the catch block I returned 200 and the pubsub message got ack:ed that everything was working (Even if it did not).

Informatica cloud (IOD) error codes

can someone please share how can I get the error codes which can cause session failure to a Informatica cloud job or please share any document/website for the same.
Thanks
You can use the activityLog REST API to get the error messages related to the task if it has failed. You can take a look at the developer guide here:
https://network.informatica.com/docs/DOC-15641
Alos, you can parse the session log available after the task run to look for error messages and codes.

One or more services have started or stopped unexpectedly SPTimerService (SPTimerV4)

I have stop and restart services(Sharepoint Administration & Sharepoint Timer Service)
I cleaned the Configuration Cache by using mentioned steps.
Summary of the steps to clear the timer job:
Stop SharePoint Timer service on all servers in the farm.
Browse to C:\ProgramData\Microsoft\SharePoint\Config{GUID} where the {GUID} folder contains a bunch of XML files and NOT the files with a “.PERSITEDFILE” extension.
Delete all the XML files
Update the contents of the Cache.ini file to just say “1” (without quotes).
Restart the SharePoint Timer service on each server
Reanalyze the issue in Health Analyzer
Does anyone know why this keeps occurring and how I can stop it?
First of all try and check your ULS Logs and see if there is any error that arise.
Secondly try and maybe check the event viewer on your SharePoint server to see if any errors are shown and make sure you have enough disk space available.
and also you might want to check this :Clearing Timer Services
Let me know if you see any error post it here.
hope it helps.
Yotam.

can you view/tail the server log rails app on travis ci

I have Rails app which has an Rspec feature with selenium that always passes locally and periodically fails on travis. It fails on click_link("my link"), with a Net::ReadTimeout: error. The stack trace isn't all that helpful and It'd be nice if there was a way to tail the log (tail -f log/test.log), so see if that's helpful...or at least view the log output. Is this possible using travis ci? I'm already waiting for ajax to finish, which suggests something external, so ultimately I'm trying to find out what request it's getting hung up on.
I believe you can cat the logs to the console as the last step of your test task, or you can use Travis' artifact option to get them uploaded to S3 - see http://docs.travis-ci.com/user/uploading-artifacts/

NATS Error while developing echo service

I'm trying to develop a system service, so I use the echo service as a test.
I developed the service by following the directions on the CF doc.
Now the echo node can be running, but the echo gateway failed with the error "echo_gateway - pid=15040 tid=9321 fid=290e ERROR -- Exiting due to NATS error: Could not connect to server on nats://localhost:4222/"
I got into this issue and struck for almost a week finally someone helped me to resolve it. The underlying problemn is something else and since errors are not trapped properly it gives a wrong message. You need to goto github and get the latest code base. The fix for this issue is http://reviews.cloudfoundry.org/#/c/8891 . Once you fix this issue, you will most likely encounter a timeout field issue. the solution for that is to define the timeout field gateway.yml
A few additional properties became required in the echo_gateway.yml.erb file - specifically, the latest were default_plan and timeout, under the service group. The properties have been added to the appropriate file in the vcap-services-sample-release repo.
Looks like the fix for the misleading error has been merged into github. I haven't updated and verified this myself just yet but the gerrit comments indicate the solution is the same as what the node base has had for some time. I did previously run into that error handling and it was far more helpful.