How to enable datasharing in Redshift cluster? - amazon-web-services

I am trying to create a datashare in Redshift by following this documentation. When I type this command:
CREATE DATASHARE datashare_name
I get this message:
ERROR: CREATE DATASHARE is not enabled.
I also tried to make it using console, but same issue.
So how to enable data sharing in Redshift ?

From the documents:here
Data sharing via datashare is only available for ra3 instance types
The document lists ra3.16xlarge, ra3.4xlarge, and ra3.xlplus instance types for producer and consumer clusters.
So, if I were in your place - I would first go back and check my instance type. If still not sure, drop a simple CS ticket and ask them if anything has changed recently & documentation is not updated

Related

What is the reason for error "Resource in project is the subject of a conflict" while trying to recreate a cloudsql instance?

I am trying to create a cloudsql instance with the following command:
gcloud beta sql instances create sql-instance-1 --tier=db-f1-micro --region=asia-south1 --network=default --storage-type=HDD --storage-size=10GB --authorized-networks=XX.XXX.XX.XX/XX
The instance sql-instance-1 is something I need not running all the time. So I create an sqldump file and when I need the database I create it. When I run this command it fails with the following error
ERROR: (gcloud.beta.sql.instances.create) Resource in project [my-project-id] is the subject of a conflict: The instance or operation is not in an appropriate state to handle the request.
From what I understand the gcloud is complaining that instance name was used before although the instance is already deleted. When I change the name to a new unused name the command works fine. The problem with this is I need to give a new name every time I re-create the instance from the dump.
My questions are:
Is this expected behavior i.e. should name of cloud-sql instance be unique and not used before within a project.
I also found that --network option is not recognized with gcloud. Seems to work only with gcloud beta as explained here. When is this expected to become GA?
This is indeed expected behaviour. From the documentation:
You cannot reuse an instance name for up to a week after you have
deleted an instance.
Regarding the --network flag and it's schedule for GA, there is no ETA for its release outside of beta. However, it's release will be listed in the Google Cloud SDK Release Notes, which you can get updates from by subscribing to the google-cloud-sdk-announce group

Issue with AWS DMS continuous replication

I am trying to create a DMS task to migrate data from RDS Postgres instance to S3 bucket. The full load works all fine, but the continuous replication is failing. Its giving this error:
"logical decoding requires wal_level >= logical"
When I checked the system settings from pg_settings, its showing that the setting "wal_level" has value "replica". So I tried to change the setting wal_level, but I am not able to find this setting in the Parameter Group in RDS. My RDS instance is using 9.6 version of parameters.
When I tried "ALTER SYSTEM SET wal_level TO 'logical'", it fails saying that "must be superuser to execute ALTER SYSTEM command", even though the user is under rds_superuser role.
Please suggest.
The Parameter name in Parameter Group is "rds.logical_replication" which needs to be changed to 1. The default value was 0.
This property changed "wal_level" value to "Logical".

AWS Data Pipeline: how to add steps other than data nodes and activities

EDIT
What I'm really asking here, is whether most folks use the "Architect" GUI to build their pipelines, or whether most folks just use JSON. Is JSON the only way to access some functionality?
/EDIT
I'm just getting started in AWS, so I'm hoping someone here can help me out.
I've used the template for "Load S3 Data into RDS MySQL table" to create a basic pipeline that does a very simple insert:
For learning purposes I want to recreate that pipeline from scratch, but I can't figure out how to add anything to the pipeline that isn't an activity or a data node. Does this have to be done through the CLI? When I try to use the "Add" button in Architect I only see options for activities and data nodes.
TaskRunners, Preconditions, Databases, Actions and Resources can be added to the Pipeline only from their respective Activities and Datanodes.
For example, An RDSDatabase can be added to the Pipeline from SqlActivity or SqlDataNode or MySqlDataNode.
Add SqlActivity -- Choose Database -- Create new: Database : Adds a Database object to the Pipeline.
Database -- Choose Type -- Select type: RDSDatabase

Trying to create a shared access policy for Azure event hub results in "Partition cannot be changed for EventHub."

I've created an Azure event hub and now I'm trying to add a "Listen" and a "Send" shared access policy. When I attempt to save them I get the following error:
SubCode=40000. PartitionCount cannot be changed for EventHub.
I'm not changing the "Partition Count", so I have not idea why I'm getting this error. Any suggestions on how to get around this problem?
I tried a similar thing with a template to update an existing Event Hub with additional consumer groups. My original template did not specify a "partitionCount" as part of the properties but when I added it the deployment worked.

How to verify that Redshift are really DISK FULL?

Question from Redshift newbies: I copy data using AWS datapipeline but it FAILED and log said
"ERROR: Disk Full Detail:
----------------------------------------------- error: Disk Full code: 1016 context: node: 0 query: 2070045 location: fdisk_api.cpp:343
process: query0_49 [pid=15048] "
I'd like to know how could we check if Redshift is really disk full via CLI or web console, any comments or hints would be appreciated.
If you're using a single node and have SQL access to the cluster (e.g. via psql), you can run:
select
sum(capacity)/1024 as capacity_gbytes,
sum(used)/1024 as used_gbytes,
(sum(capacity) - sum(used))/1024 as free_gbytes
from
stv_partitions where part_begin=0;
This articles has more: https://www.flydata.com/blog/querying-free-disk-space-on-redshift/
You can check that in CloudWatch console. In the left bar, you'll see bunch of AWS services under the 'Metrics' heading. Click on Redshift. In that, look for the 'PercentageDiskSpaceUsed' metric for the concerned cluster.
Also, do remember that this metric is separately available for each compute node.