Thanks in advance.
I'm using IBM Data Science Experience (DSX). In order to use Decision Optimization in DSX, I need to get a DOcplexcloud API Key.
How to get the API Key?
Once again, thank you very much for the time you've given.
Regards,
Deb
You need to either need to have a Docplex Cloud Website account
https://dropsolve-oaas.docloud.ibmcloud.com/software/analytics/docloud
and follow this document to get the API Key.
https://developer.ibm.com/docloud/documentation/decision-optimization-on-cloud/api-key/
or
You can provision Decision Optimization Instance on bluemix
https://console.bluemix.net/catalog/services/decision-optimization?env_id=ibm%3Ayp%3Aus-south
{
"url": "https://api-oaas.docloud.ibmcloud.com/job_manager/rest/v1",
"client_id": "bm_api_ext_4ea008b2-687d-4ce4-8b0a-712f0ed680e1",
"subscriptionId": "bm_45b16939-2382-4f6f-a2d2-c9ae5aae87c1"
}
When you using it on DSX,
With Python API,(https://developer.ibm.com/docloud/documentation/docloud/python-api/python-example/)
use url for BaseURL
and client_id is API Key.
https://developer.ibm.com/docloud/documentation/docloud/rest-api/rest-api-example/
Here is the whole notebook:-
https://github.com/charles2588/bluemixsparknotebooks/blob/master/Python/docplex.ipynb
Related
This is more of a conceptual question as there is no relevant documentation available. We have an on prem IBM-MQ from which we need to transfer data on our cloud storage bucket (GCP/AWS), what could be possible solutions in this case? Any help or direction would be appreciated. Thank you!
I'm assuming you can reach your goal once the MQ-data has been changed/converted to supported format by the Big Query.
You can refer on this google documentation for full guide on Loading data from local files. You can upload file via GCP Console or using selected programming language that will match on your on-prem. There's also variety of uploads that you can choose from according to data file. This also includes the right permission to use the BigQuery.
If you require authentication you check on this Big Query Authentication Guide
I am working with my team to prep a project for a potential client. We've researched Amazon MWS API, and we're trying to develop an algorithm using the data scraped from this API.
Just want to make sure we understand the research correctly:
Is it possible to scrape data from Amazon.com like the plugins RevSeller or HowMany do? Then can we add that data to a database for use in an algorithm to determine whether or not an Amazon reseller should invest in reselling a product?
Thanks!
I am doing a similar project. I don't know the specifics of RevSeller or HowMany, but another very popular plugin is Amzpecty. If you use a tool like Fiddler, you can see the HTTP traffic and figure out what it does. They basically scrape out the ASIN and offer listing ID's on the current page you are looking at and one-by-one call the Amazon Product Advertising API, which is not the same thing as MWS. Out of that data returned, they produce a nice overlay that tells you all kinds of important stuff.
Instead of a browser plugin, I'm just writing an app that makes HTTP calls based on a list of ASIN's to the PA API and then I can run the results through my own algorithms. Hope that gives you a starting point.
This is just a general question, but I'm wondering if there's an existing API that displays the current price for an item on Amazon? As in, if the price changes, the site will reflect that change as well.
If not, would building a web crawler to go through and find the Amazon items of my choice be the best way to build my own version of this? If so, what language would you recommend to begin this sort of project.
I'm not sure if I should have actually asked this in SuperUser but I appreciate the input. Thanks guys!
There are plenty of web crawling services for this task.
https://import.io/
https://www.kimonolabs.com/
http://www.diffbot.com/
If you want to make your own, I recommend node.js because it's asynchronous behavior.
Amazon Mechanical Turk is a mass-micro outsourcing API, where you can get lots's of small simple tasks (e.g."Is there a shop in this image") done relatively cheaply (e.g. 0.10 U$ per image).
Amazon seems to assume that this service is mostly of interest to US companies. This results in difficulties if you want to use the service (as an "Requester"): for example there is no easy way of funding your Account without an US Bank account.
Can somebody share experiences of using the Service from outside the US?
Since Amazon doesn't let outside US access yet, you'll need to either:
Wait for Amazon to open it up to outside US, might be a long wait.
Get a good friend in the US to open a PO Box for you then use an online US bank to create an account. If you don't start a corporation or other legal business, they will need to provide their US drivers license to Amazon too.
Use a 3rd party provider like HIT-Builder who will let you post on Mechanical Turk using their US Amazon account.
I think they recently changed their rules - I am based in Australia, and as recently as 6 months ago I was unable to use Mechanical Turk. That recently changed and now I am able to post jobs on the service.
You may want to take a look at TurkPrime labs which offers this as a service to non-US based researchers.
The problem you have is that to actually be able to add funds to your account, you need to have a US billing address. If you can get round that issue, then it all works fine.
In dopplr [http://www.dopplr.com] there is an option to fill the city of travel and the site will automatically find the city around the world. Is there any web service or database for such a city lookup?
http://www.geonames.org/ is useful - HTML query, download and web service.
Also http://developer.yahoo.com/geo/ - you'll need a Yahoo developer ID.
Google maps API includes a geocoding service and there are plenty of other options
You could download and process this list on wikipedia. If you trust it.
The best free database of world cities I have found is from MaxMind. See http://www.maxmind.com/app/geolitecity It also includes geocoding information and population. Uses ISO standard for country and regions.
There's the OpenStreetMap Name Finder.
I don't know if it's necessarily the most complete resource, but AFAIK OSM is free for any use, which may/may not be the case with other options.