where i can get the postman collection data in windows 10? - postman

To clarify, I want to find where Postman saves collection files to by default when online syncing is disabled
I've looked in %LocalAppData%, My Documents, and Program Files, but I don't see where Postman saves its collection data

C:\Users\your username\AppData\Roaming\Postman\IndexedDB\

you have to get the details form Roaming\Postman not in LocalAppData

Postman collection data can be found in several places in Windows 10. Here are a few possible options:
In the Postman application: If you have Postman installed on your Windows 10 machine, you can find your collection data by opening the application and going to the "Collections" tab in the left-side navigation panel.
In the Postman data folder: Postman stores its data, including collections, in a specific folder on your computer. On Windows 10, the default location for this folder is %APPDATA%\Postman. You can access this folder by opening the File Explorer and pasting the above path in the address bar.
In the Postman cloud: If you have a Postman account and have synced your collections to the cloud, you can access your collection data by logging into your Postman account at https://app.getpostman.com/.
Regardless of where you find your collection data, you can export it from Postman by clicking the "Export" button in the top right corner of the Collections tab. This will allow you to save the collection data as a JSON file that can be imported into Postman or other tools.

Related

Access Webpage With Credentials and Cookies From Command Line

I am trying to access a proprietary website which provides access to a large database. The database is quite large (many billions of entries). Each entry in the database is a link to a webpage that is essentially a flat file containing the information that I need.
I have about 2000 entries from the database and their corresponding webpages in the database. I have two related issues that I am trying to resolve:
How to get wget (or any other similar program) to read cookie data. I downloaded my cookies from google chrome (using: https://chrome.google.com/webstore/detail/cookiestxt/njabckikapfpffapmjgojcnbfjonfjfg?hl=en) but for some reason the html downloaded by wget still cannot be rendered as a webpage. Similarly, I have not been able to get Google Chrome from the command line to read cookies. These cookies are needed to access the database, since they contain my credentials.
In my context, it would be OK if the webpage was downloaded as a PDF, but I cannot seem to figure out how to download a webpage as a pdf using wget or similar tools. I tried using automate-save-page-as (https://github.com/abiyani/automate-save-page-as) but I continuously get an error of the browser not being in my PATH.
I solved both of these issues:
Problem 1: I switched away from wget, curl and python's requests to simply using the selenium webdriver in python. Using selenium, I did not have to deal with issues such as passing cookies,headers, post and get, since it actually opens a browser. This also has a plus that as I was writing the script to use selenium, I could inspect the page and see what it was doing as it was doing it.
Problem 2: Selenium has a method called page_source, which downloaded the html of the webpage. When I tested it, it rendered the html correctly.

WSO2 EMM - Manage data to wrapping them in multi page

I have added new report to EMM, but as the report data recorde are too much, I would like to wrap them in multiple pages like configuration tab (users, roles and policies), How can I fix this or how it works for those tabs?
Thanks
WSO2 EMM is built on top of bootstrap 2. You can add Bootstrap tabs as mention in following doc and have multiple tabs on your page
http://getbootstrap.com/2.3.2/javascript.html#tabs
Edit:
WSO2 EMM 1.1.0 is built on top of a Jaggery MVC framework.
Create a report by adding a function to emm/controller/reports.js. This is a controller and you can use modules to retrieve data to the controller. see devices_complience function for example in the following file
https://github.com/wso2/product-emm/blob/master/modules/apps/emm/controller/reports.js
Add the corresponding view file to emm/views/reports
Eg: https://github.com/wso2/product-emm/blob/master/modules/apps/emm/views/reports/devices_complience.hbs
Add a link to your report in the emm/views/reports/index.hbs file
If you need to have additional java-script functions you can add them to emm/client/js/reports/reports.js file
I should edit "Carbone_Home\repository\deployment\server\jaggeryapps\emm\client\js\reports.js" file to managing record on table.
"sDom":"t"
to
"sDom":"<'row-fluid't<'row-fluid'<'span6'i><'span6'p>>"
and for setting number of record per page I should set this :
"iDisplayLength": 10,
which uses emm\assets\jqueryext\datatables\jquery.datatables.

Azure Web Job writes files in a website's directory

I attached a Web Job to my Azure website. The webjob prepares a file and I want to save it on a proper folder in the website.
Environment.CurrentDirectory run on the script returns a path under a Temp directory: Temp\jobs\triggered\WEBJOBNAME\q0uwrohv.x5e
I tried to go down on the directory tree:
string path = Path.Combine(Environment.CurrentDirectory, #"..\..\..\..\..\Data")
But it doesn't work:
C:\DWASFiles\Sites\WEBSITENAME\Temp\jobs\triggered\WEBJOBNAME\q0uwrohv.x5e\..\..\..\..\..\Data
How to make and save files from WebJob to a particular path?
I don't want to use blob store.
The path for the root of your Azure Web Site is (usually) d:\home\site\wwwroot.
d:\home is also stored in an environment setting called %HOME%.
To get more insight on the different paths you can use on your site go to:
https://{sitename}.scm.azurewebsites.net, there' you'll have the Debug Console where you can browse through your site and Environment to see all the environment variables you can use.
Your WebJob will have access to the same paths/environment as your Web Site.
For more information on this administration site go to:
http://azure.microsoft.com/blog/2014/03/28/windows-azure-websites-online-tools-you-should-know-about-2/
try the following instead of putting a file location in the value parameter just put this as I show you here
you can do it in the app.config file
add key="app:TempFolderPath" value="~/temp/"/
add key="app:TempReportDirectory" value="~/temp/"/
the web job will automatically put in this location of
D:\local\Temp\jobs\continuous\ImporterWebJob\yex3ad1c.3wo\~\temp\...your file...
I hope this will not give you any errors.

python upload image from a url to google drive api

For my Python app,I had completed the basic settings to interact with google drive api and found it working by a test upload of a CSV file. Now I need to upload an image from a url to a newly created folder named 'myappname' in Google Drive.
Thanks in advance
For now, there is no way you can directly upload file from url. There are two workaround I can think of
Download file and upload it back using Files.insert()
Use Save to Drive button
Using save to Drive button requires user interaction to click the button which might not be the one you want. In that case, downloading and uploading is the only way I can think of.

Tracking number of downloads in Django

I have a website where I allow people to download files, such as Powerpoint templates that they can reuse.
I added a column to the model that stores information and location of these files called 'Downloads' and in that column I'd like to use it as a counter to track how many times the file has been downloaded.
How would I go about this? Would it be a GET request on the page? Basically I just provide a link to the file and the download starts automatically, so not sure how to relay the fact that the link was clicked back to the model.
Many thanks :-)
If you create a view that handles the GET request you could put the updating code in that view. If your Django-app is not handling the uploading itself, you can create a view that simply redirects to the download link, after updating the counter in the database.
You have several ways to do this:
create a custom view, that "proxies" all files, while keeping track of downloads
create a middleware that does pretty much the same as above, filtering which requests to record
..but none of the above will be applicable if you want to count downloads of collected static files, that will be served directly by your http server, without passing through django. In this case, I'd retrieve the downloads count from the webserver logs; have a look if your webserver allows storing logs to database, otherwise I'd create a cron-running scripts that parses the logfiles and stores the count to a db, accessible from your django application.
Like redShadow said you can create a proxie view. This view could serve the files via mod_xsendfile (if you are using apache as webserver) and set a counter for the downloads.