Difference between JDBC driver vs Python Adapters for Amazon Redshift - amazon-web-services

To connect Amazon Redshift, I used the python psycopg2 module to inject the dataset on Amazon Redshift and it is working fine. Just to mention that I'm using the Redshift's endpoint URL to connect via the psycopg2 which underneath uses python DB API v2.0. Amazon Redshift also provides the JDBC or ODBC URL to connect but I'm not using this.
pseudo-code:
import psycopg2
try:
connection = psycopg2.connect(user = "redshift_user",
password = "redshift_password",
host = "redshift_endpoint",
port = "5432",
database = "redshift_database")
cursor = connection.cursor()
Note: I also use the same endpoint of Redshift to connect it from my different clients like Tableau, Navicat Premium or other SQL clients that use JDBC/ODBC driver underneath it.
But recently Amazon sent me to message about JDBC driver update,
AWS Redshift identified an issue in Redshift JDBC drivers that led to
unexpected server restarts, which was subsequently fixed in the latest
Redshift JDBC drivers. Some of your clusters in the US-WEST-2 Region
are registering connections from older versions of the JDBC driver and
might be affected by this issue.
Please upgrade your driver to the latest version: 1.2.36.1060, which
is available for download [1].
[1] https://docs.aws.amazon.com/redshift/latest/mgmt/configure-jdbc-connection.html#download-jdbc-driver
So now I have a couple of question-
The psycopyg2 module is using the JDBC or ODBC driver underneath it?
The above message from AWS is because of the different clients I used that underneath is using the older version of JDBC or ODBC drivers. So I've to update the drivers of my clients only not on my psycopg2 module.

Related

PowerBI Pro unable to test connection to data source, Redshift

Was able to connect PowerBI Desktop to a Redshift Cluster and visualize data from there.
However, I have a need to setup Autorefresh and other options that require the Data source credentials to be maintained and connected.
Importing my data from PowerBI Desktop to Pro (web version), it is unable to connect.
On PowerBI Pro it is unable to make such connections and receiving the following. Also verified that my credentials are correct.
Any help would be appreciated.

How to use Presto with Power BI

Has anyone tried to connect power BI tool with Presto?
I have used the connectors and driver available on the Qubole documentation site. I have installed the drivers and connectors but it is throwing error on accessing S3 location.
Qubole drivers and connector will work only with Qubole's Managed Presto Service.
Disclaimer: I work for Qubole.

Can SQL Developer be used with the Athena JDBC Driver

I'm trying to connect to Athena using the JDBC drivers provided by Amazon, and using SQL Developer as the client. So far, I haven't had any luck with Java 1.8.181 and AthenaJDBC42-2.0.7.jar Has anyone had any luck on this front? Before I try mixing up which versions of Java, JDBC driver, and/or SQL Developer, I thought I'd at least ask if anyone has been successful using SQL Developer with the Athena JDBC drivers.
No.
SQL Developer doesn't allow for just any JDBC driver to be added...we restrict connectivity to the platforms we officially support for database migrations to the Oracle Database platform.
Athena doesn't have migration support, hence the lack of connectivity. If you need assistance with a migration, please send me a note.

Can SourcePro DB connect to Oracle using PKI?

Currently I am using SourcePro DB to connect to oracle database using id, password and SID via DB Access Module for Oracle OCI. However I will need to use Oracle PKI in the future. So wondering if SourcePro DB supports Oracle PKI and if anyone has done that before.
We checked with rogue wave and currently SourcePro doesn't support DB connection using Oracle PKI.

WSO2 API Manager with Mysql Cluster backend - NDB sql scripts

I'm setting up the WSO2 API Manager with a MySQL Cluster backend. In the /dbscripts directory within the install directory for the API Manager, there are .sql scripts for setting up the database with either the mysql.sql script, which uses InnoDB, or the mysql_cluster.sql script which uses NDB. My problem is the apimgt database, which from this article: http://docs.wso2.org/display/Cluster/Clustering+API+Manager , says to use the script in /dbscripts/apimgt/mysql.sql, but there is no equivalent script for the MySQL Cluster setup. I attempted to just replace all instances of ENGINE InnoDB with ENGINE NDB in the script, but when trying to import that into MySQL I get: ERROR 1215 (HY000) at line 82: Cannot add foreign key constraint. Does anyone know of a MySQL Cluster script for this database?
Thanks.
API Manager does not provide mysql cluster script with the pack.
For converting it to use in mysql cluster,
Yes you need to change the engine to NDB.
Also you have to remove all the forign key constraints and indexes.