Google big query data server connection with cognos Analytics 1116 - google-cloud-platform

I am trying to create google big query data server connection with cognos analytics on cognos portal but i am getting below error . I have followed all steps . Placed json in the driver folder inside cognos install directory along with all the simba jdbc driver but not sure where i am missing .I have installed simba odbc drivers as well , system dsn is also showing simba driver but still getting connection refused. My cognos setup is installed on windows server 2019 data center which is Virtual Machine hosted on GCP and port 80,443 traffic is also enabled.
XQE-DS-0006 Unable to logon to the data source.
[Simba]JDBC Connection Refused: [Simba]JDBC Required Connection Key(s): OAuthServiceAcctEmail; [Simba]JDBC Optional Connection Key(s): AdditionalProjects, AllowLargeResults, DefaultDataset, EnableHighThroughPutAPI, FetchThreadsPerResultSet, ForceHTAPI, HighThroughPutMinTableSize, IgnoreTransactions, KMSKeyName, Language, LargeResultDataset, LargeResultsDatasetExpirationTime, LargeResultTable, MaxResults, OAuthAccessToken, OAuthClientId, OAuthClientSecret, OAuthRefreshToken, ProxyHost, RequestGoogleDriveScope, StringColumnLength, Timeout, TimestampFallback, useQueryCache

Please see the following document which goes through the connection step by step.
Google big query data server connection with cognos Analytics 1116
Its most likely the JSON file is not in the JDBC connection URL.

Related

ODBC connection gives Unknown Host error from Power BI

I am using an ODBC driver of MarkLogic instance which I am using in Power BI to access MarkLogic database files.
I am using 64 bit ODBC driver and have set up the connection and tested the connection successfully. I have also accessed the MarkLogic files under Documents Database.
But today when I tried to connect from Power BI, it gives me this error:
Could not translate hostname "myhostname" to address : Unknown host
The same error is thrown when I am trying to test the connection in ODBC set up.
The driver used is Marklogic SQL.
MarkLogic is up and running, and there have been no changes done in any connection properties so far.

connecting PBI to impala

I created a cloudera cluster (ENTERPRISE DATA HUB) on azure. I can use the DNSname:7180 to view and manage cluster. However, I am not successfull in connecting to the Impala from PowerBI Desktop. I tried both VM names with dn0 and mn0 extension ([myhostname]-dn0.eastus2.cloudapp.azure.com) and ports 71890, 21000, 21050, based on this and this
It always fail at the authentication level. I tried anonymous, windows and DB authentication, and they all failed with this error:
Details: "ODBC: ERROR [08S01] [Microsoft][ImpalaODBC] (100) Error from the Impala Thrift API: connect() failed: errno = 10060"
Any help or clue is appreciated.
The port is 21050. You have to open it on Azure VM since it is not open by default.

Pentaho DI can't connect to AWS Redshift - Amazon Error 100021

Referring to Pentaho's Doc, we should be using RedshiftJDBC4.jar instead of version 4.1. I have downloaded the driver and placed it in the lib/ directory. Relaunched spoon.sh and I noticed it is no longer complaining about not able to find the com.amazon.redshift.jdbc4 class driver as I was using the 4.1 driver earlier. However, it still could not establish the connection.
Error connecting to database [aws_redshift] :
org.pentaho.di.core.exception.KettleDatabaseException: Error occurred
while trying to connect to the database
Error connecting to database: (using class
com.amazon.redshift.jdbc4.Driver) Amazon Error setting
default driver property values.
Can anyone help on this?
On the flip side, I can connect to my endpoint using SQLWorkbench/J, a SQL client tool.
Somehow I managed to get it working. It seems that downloading AWS Redshift drivers version 4, 4.1, or 4.2 and placing them in the lib/ directory did not work for me for each version by choosing Redshift as connection type (in Database Connection setup).
Instead, I chose PostgreSQL using JDBC. In host name field, I included the endpoint WITHOUT port number 5439 at the end. So, the endpoint should just end with ...amazonaws.com. Fill in database name, port number of 5439, and username and password. If this did not work, try downloading the latest PostgreSQL JDBC driver and placing it in lib/ directory and try again.

Test connection failed in Informatica Developer

I am working with Informatica Developer - Power-Center. I have it installed on 2 computers: my local PC and the MS Server where the Informatica Server is also installed.
I defined a new connection and tested it by clicking Test Connection. On my PC the test works fine but when I test the connection on the server the test fails with the following error: "[REL_10611] Error returned during ODBC access. ErrorCode=[-1]. Reason=[[Microsoft][ODBC Driver Manager] Data source name not found and no default driver specified]."
Another strange thing is that the mapping that includes a relational data object that uses this connection works even on the server and adds row to the DB.
Does anybody know why this could be happening? Is this a bug in the Informatica Developer?
Did you create the ODBC data source on the server? The data source needs to be created on both your local PC and on the server. When using the Developer Tool (and the Power Center client tools for that matter), some operations are executed from your local workstation and some are executed on the Data Integration Service (DIS).

ColdFusion DSN with DB2 via ODBC

I'm attempting to connect a ColdFusion application to a DB2 ODBC DSN.
Here's my error message:
Connection verification failed for data source: <DSN NAME>
java.sql.SQLException: [Macromedia][SequeLink JDBC Driver][ODBC Socket][IBM][CLI Driver] SQL30082N Attempt to establish connection failed with security reason "24" ("USERNAME AND/OR PASSWORD INVALID"). SQLSTATE=08001
The root cause was that: java.sql.SQLException: [Macromedia][SequeLink JDBC Driver][ODBC Socket][IBM][CLI Driver] SQL30082N Attempt to establish connection failed with security reason "24" ("USERNAME AND/OR PASSWORD INVALID"). SQLSTATE=08001
I've installed DB2 client tools on the server ColdFusion runs on. I've verified the credentials are correct.
I'm not a DB2 guy, but have you seen the Coldfusion DB2 Universal Driver doc located here?
http://livedocs.adobe.com/coldfusion/8/htmldocs/help.html?content=datasources_ADV_MJS_07.html
Based on your comments, you're using ODBC Socket instead. So while this isn't 1 to 1 what you're asking, it might have something useful: Configure Solaris to DB2 ODBC
http://kb2.adobe.com/cps/171/tn_17188.html
The other thing to keep in mind is that DB2 is only supported in CF Enterprise and Developer Editions. http://www.adobe.com/products/coldfusion/systemreqs/
http://kb2.adobe.com/cps/801/80121c8.html
This CFMX doc ended up being what we went with to get it working.
Thanks to SO for pointing us in the right direction.