When I'm trying to connect Snowflake to Power BI I get an error:
"ODBC: ERROR [HY000] [Microsoft][Snowflake] (4)
REST request for URL https://eda87722.snowflakecomputing.com:443/session/v1/login-request?requestId=9eb99320-1fc6-49ee-859b-0fca3e70638b&request_guid=5f7d40b7-7025-410c-8905-da5b7e7e78cd&warehouse=COMPUTE_WH failed: CURLerror (curl_easy_perform() failed) - code=5 msg='Couldn't resolve proxy name' osCode=9 osMsg='Bad file descriptor'."
Interesting that I'm able to establish the connection on another machine with no issues. But I can't find the reason for such issue.
What I've tried:
Established the connection with vpn on and vpn off - no difference
Checked environmental variables http_proxy and https_proxy (http://proxyserver.internal) - they exist
Checked my access to data in Snowflake - I have access
Reinstalled Power BI
Tried to install Snowflake ODBC driver, but it is not visible in the list of my drivers
Tried to connect to different warehouses in Snowflake - same issue
'Auto resume' option for my warehouse in on
Could you, please, advise what else can I do?
Thank you.
Related
I am using dataverse to extract data from CRM into Power BI Power Query. For the past few weeks, there hasn't been any issues with establishing a connection and getting the data. However, just earlier this week, I have been unable to retrieve any sort of data; instead, I encounter the following error -
DataSource.Error: Microsoft SQL: A connection was successfully established with the server, but then an error occurred during the pre-login handshake. (provider: SSL Provider, error: 0 - The wait operation timed out.)
Details:
DataSourceKind=CommonDataService
DataSourcePath=organization CRM web portal here
Message=A connection was successfully established with the server, but then an error occurred during the pre-login handshake. (provider: SSL Provider, error: 0 - The wait operation timed out.)
ErrorCode=-2146232060
Number=258
Class=20
As far as I know, there has not been any changes to my user privileges, or any changes to our servers. Furthermore, by using my personal mobile hotspot, I was able to successfully retrieve all the data from CRM. However, when I swapped back to my office's WiFi, the error occurred again. Is there any way to troubleshoot or check what is the cause of the problem (which is likely my office's WiFi/Network)?
I'm wanting to create a one-way real-time copy of a Salesforce (SF) object in Redshift. The idea being that when fields are updated in SF, those fields will be updated in Redshift as well. The history of changes are irrelevant in AWS/Redshift, that's all being tracked in SF - I just need a real-time read-only copy of that particular object to query. Preferably without having to query the whole SF object, clearing the Redshift table, and piping the data in.
I thought AWS AppFlow listening for SF Change Data Capture events might be a good setup for this:
When I try to create a flow, I don't have any issues with the SF source connection:
so I click "Connect" in the Destination details section to setup Redshift and I fill out this page and click "Connect" again:
About 5 seconds goes by and I receive this error pop-up:
An error occurred while creating the connection
Error while communicating to connector: Failed to validate Connection while attempting "select distinct(table_schema) from information_schema.tables limit 1" with connector failure Can't connect to JDBC database with message: Amazon Error setting/closing connection: SocketTimeoutException. (Service: null; Status Code: 400; Error Code: Client; Request ID: null; Proxy: null)
I know my connection string, username, password, etc are all good - I'm connected to Redshift in other apps. Any idea what the issue could be? Is this even the right solution for what I'm trying to do?
I solved this by adding the AppFlow IP ranges for my region to my Redshift VPC's security group inbound rules.
Following the steps in https://cloud.google.com/bigquery/docs/cloud-sql-federated-queries#setting_up_database_connections after the API was activated, I tried to set up a connection between our BigQuery database and our Postgresql CloudSQL database. I entered everything correctly:
When I try to connect it, I get the following error, without any error code or other information as to why it might have went wrong:
Getting this error when connecting Power BI with Azure Databricks through spark build in connector:-
Details: "ODBC: ERROR [HY000] [Microsoft][DriverSupport] (1170)
Unexpected response received from server. Please ensure the server
host and port specified for the connection are correct."
I have checked many times host and port of the databrick cluster , and also tried after restarting of cluster .
Guide for the connection:-
https://docs.azuredatabricks.net/user-guide/bi/power-bi.html
Got the same problem today. I followed these instructions and it worked.
The user was not able to import SQL data Power BI and getting this error, while testing connection in ODBC was successful.
It turned out that he has old credentials stored in PowerBI, and that caused identification issues. Purging cached data sources (Power BI: Home >Edit Queries > Data source settings" resolved the issue.
I am trying to connect to Athena using SQL workbench. I followed all the instructions from page 15 to 19 mentioned in this PDF file:
https://s3.amazonaws.com/athena-downloads/drivers/JDBC/SimbaAthenaJDBC_2.0.7/docs/Simba+Athena+JDBC+Driver+Install+and+Configuration+Guide.pdf
If I use the default athena bucket name, I get this error:
S3://aws-athena-query-results-51346970XXXX-us-east-1/Unsaved
[Simba]AthenaJDBC An error has been thrown from the AWS SDK
client. Unable to execute HTTP request: No such host is known
(athena.useast-1.amazonaws.com) [Execution ID not available]
For any other bucketname I get this error:
s3://todel162/testfolder-1
[Simba]AthenaJDBC An error has been thrown from the AWS SDK
client. Unable to execute HTTP request: athena.useast-1.amazonaws.com
[Execution ID not available]
How do I connect to Athena using JDBC client?
Using copy-paste had an issue with the string on page 16:
jdbc:awsathena://AwsRegion=useast-1;
It should have a - like this...
jdbc:awsathena://AwsRegion=us-east-1;
Once I corrected this, I was able to connect.