I created a cloudera cluster (ENTERPRISE DATA HUB) on azure. I can use the DNSname:7180 to view and manage cluster. However, I am not successfull in connecting to the Impala from PowerBI Desktop. I tried both VM names with dn0 and mn0 extension ([myhostname]-dn0.eastus2.cloudapp.azure.com) and ports 71890, 21000, 21050, based on this and this
It always fail at the authentication level. I tried anonymous, windows and DB authentication, and they all failed with this error:
Details: "ODBC: ERROR [08S01] [Microsoft][ImpalaODBC] (100) Error from the Impala Thrift API: connect() failed: errno = 10060"
Any help or clue is appreciated.
The port is 21050. You have to open it on Azure VM since it is not open by default.
Related
I am trying to set up a connection to a postgres DB from Cloud Fusion. I have what I think is the appropriate JDBC driver. I am going through the Wrangler UI. I keep getting this error message: Failed to explore connection. Error: "Failed to create connection to database via connection string: jdbc:postgresql://XX.XXX.X.XX:PORT/TABLE_NAME and arguments: {user=USERNAME}. Error: SocketTimeoutException: connect timed out."
It appears my password is being sent. Is this a jdbc issues or something else? All help appreciated.
Thanks!
I am trying to create google big query data server connection with cognos analytics on cognos portal but i am getting below error . I have followed all steps . Placed json in the driver folder inside cognos install directory along with all the simba jdbc driver but not sure where i am missing .I have installed simba odbc drivers as well , system dsn is also showing simba driver but still getting connection refused. My cognos setup is installed on windows server 2019 data center which is Virtual Machine hosted on GCP and port 80,443 traffic is also enabled.
XQE-DS-0006 Unable to logon to the data source.
[Simba]JDBC Connection Refused: [Simba]JDBC Required Connection Key(s): OAuthServiceAcctEmail; [Simba]JDBC Optional Connection Key(s): AdditionalProjects, AllowLargeResults, DefaultDataset, EnableHighThroughPutAPI, FetchThreadsPerResultSet, ForceHTAPI, HighThroughPutMinTableSize, IgnoreTransactions, KMSKeyName, Language, LargeResultDataset, LargeResultsDatasetExpirationTime, LargeResultTable, MaxResults, OAuthAccessToken, OAuthClientId, OAuthClientSecret, OAuthRefreshToken, ProxyHost, RequestGoogleDriveScope, StringColumnLength, Timeout, TimestampFallback, useQueryCache
Please see the following document which goes through the connection step by step.
Google big query data server connection with cognos Analytics 1116
Its most likely the JSON file is not in the JDBC connection URL.
I am attempting to take my existing cloud composer environment and connect to a remote SQL database (Azure SQL). I've been banging at my head at this for a few days and I'm hoping someone can point out where my problem lies.
Following the documentation found here I've spun up a GKE Service and SQL Proxy workload. I then created a new airflow connection as show here using the full name of the service azure-sqlproxy-service:
I test run one of my DAG tasks and get the following:
Unable to connect: Adaptive Server is unavailable or does not exist
Not sure on the issue I decide to remote directly into one of the workers, whitelist that IP on the remote DB firewall, and try to connect to the server. With no command line MSSQL client installed I launch python on the worker and attempt to connect to the database with the following:
connection = pymssql.connect(host='database.url.net',user='sa',password='password',database='database')
From which I get the same error above with both the Service and the remote IP entered in as host. Even ignoring the service/proxy shouldn't this airflow worker be able to reach the remote database? I can ping websites but checking the remote logs the DB doesn't show any failed logins. With the generic error and not many ideas on what to do next I'm stuck. A few google results have suggested switching libraries but I'm not quite sure how, or if I even need to, within airflow.
What troubleshooting steps could I take next to get at least a single worker communicating to the DB before moving on the the service/proxy?
After much pain I've found that Cloud composer uses ubuntu 1804 which currently breaks pymssql as per here:
https://github.com/pymssql/pymssql/issues/687
I tried downgrading to 2.1.4 to no success. Needing to get this done I've followed the instructions outlined in this post to use pyodbc.
Google Composer- How do I install Microsoft SQL Server ODBC drivers on environments
I have downloaded Talend Open Studio v5.6 Big Data Version and Hortonworks Sandbox 2.1 and Cloudera Quickstart VM 5.2 and able to interact with and upload data to both HDP and CDH through their Hue interface. But, when I am trying to use the Talend components to make data transfer to HDFS from an Oracle RDBMS, I am getting stuck at different phases which I am listing below.
I am using VM manager as Oracle Virtual Box and could only use NAT as network configuration as other configurations are disallowed in my organization for security reasons.
Exhibit 1: Using tSqoop/tHDFS component to transfer data to HDFS{Hortonworks Sandbox 2.1}
I am using below configuration
NameNode URI: "hdfs://127.0.0.1:8020"
ResourceManager: "127.0.0.1:8032"*
Error with tSqoop:
[ERROR]: org.apache.sqoop.tool.ImportTool - Encountered IOException running import job: java.io.IOException: Failed on local exception: java.io.IOException: An existing connection was forcibly closed by the remote host; Host Details : local host is: "PTP-A4A-020/10.206.8.142"; destination host is: "127.0.0.1":8032;
Error with tHDFS:
Exception in component tHDFSOutput_1
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /user/hue/sqoop-oracle/SCD_POC/STG_OVERD_LN.txt could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are excluded in this operation.
...
Exhibit 2: Using tHive component to interact with Hive tables {Cloudera Quickstart VM 5.2}
I am using hostname:"127.0.0.1" and port:"10000" and User:"hue"
Error:
Exception in component tHiveConnection_1
java.sql.SQLException: Could not establish connection to 127.0.0.1:10000/default: java.net.ConnectException: Connection refused:
However, when using the above NameNode and ResourceManager URI configurations, I am able to load data into HDFS {Cloudera} with tSqoopImport.
I have it worked using "Host Only Adapter" configuration on virtual box -> Network. Basically it assigned a new IP address (type Class C) within default 10.0.2.15 and thus I can have my application (talend) communicate to NameNode or DataNode present in Cloudera VM.
I'm attempting to connect a ColdFusion application to a DB2 ODBC DSN.
Here's my error message:
Connection verification failed for data source: <DSN NAME>
java.sql.SQLException: [Macromedia][SequeLink JDBC Driver][ODBC Socket][IBM][CLI Driver] SQL30082N Attempt to establish connection failed with security reason "24" ("USERNAME AND/OR PASSWORD INVALID"). SQLSTATE=08001
The root cause was that: java.sql.SQLException: [Macromedia][SequeLink JDBC Driver][ODBC Socket][IBM][CLI Driver] SQL30082N Attempt to establish connection failed with security reason "24" ("USERNAME AND/OR PASSWORD INVALID"). SQLSTATE=08001
I've installed DB2 client tools on the server ColdFusion runs on. I've verified the credentials are correct.
I'm not a DB2 guy, but have you seen the Coldfusion DB2 Universal Driver doc located here?
http://livedocs.adobe.com/coldfusion/8/htmldocs/help.html?content=datasources_ADV_MJS_07.html
Based on your comments, you're using ODBC Socket instead. So while this isn't 1 to 1 what you're asking, it might have something useful: Configure Solaris to DB2 ODBC
http://kb2.adobe.com/cps/171/tn_17188.html
The other thing to keep in mind is that DB2 is only supported in CF Enterprise and Developer Editions. http://www.adobe.com/products/coldfusion/systemreqs/
http://kb2.adobe.com/cps/801/80121c8.html
This CFMX doc ended up being what we went with to get it working.
Thanks to SO for pointing us in the right direction.