I do not know why AWS DocumentDb does not support MongoDB version that above 3.6? Should I use mongo 3.6 or 4.0 above?
Amazon DocumentDB now supports MongoDB 4.0 compatibility including transactions: https://aws.amazon.com/about-aws/whats-new/2020/11/amazon-documentdb-with-mongodb-compatibility-adds-support-for-mongodb-4-and-transactions/
Document DB is compatible with only MongoDB 3.6.
See: https://aws.amazon.com/documentdb/features/
Whether you want to use 3.6 vs 4.0 or even 4.2 or 4.4 is very subjective to what you want to do with the DB.
The pros of using Document DB is that its a managed service and hence you dont have to worry a lot about setting it up.
The con is you will not get features present in version 3.6 onwards example multi document transactions, new operators in the agg pipeline, bug fixes etc.
To figure out the exact changes check - https://docs.mongodb.com/manual/release-notes/
You can install mongodb on EC2 instances, this will ensure you will get the latest and greatest mongodb. However this comes with the added work of managing the mongodb instance, its backups, High availability considerations etc.
Do note: No matter what you decide I would recommend that you try to use the latest drivers that are present today so that you have the freedom to go to the latest version of self installed mongo or even upgrade document db engine versions as they become available.
AWS DocumentDb uses its own database engine compatible with MongoDB 3.6 API.
MongoDB 4.0 API is not supported as of yet.
We have MongoDB Atlas available in AWS - AWS also supports fully managed database service - MongoDB Atlas - https://aws.amazon.com/de/quickstart/architecture/mongodb/
AWS enables to set up the infrastructure to support MongoDB deployment in a flexible, scalable, and cost-effective manner on the AWS Cloud.
https://aws.amazon.com/de/quickstart/architecture/mongodb/
Related
I am trying to collect logs/metrics from kafka connector in GCP.
I read the documentation
It seems that ops agent only support till Apache Kafka Version 3.0.0 and I am using Kafka version 3.2.3.
May I know if there is any roadmap of version support of Kafka in ops agent?
Or any alternatives I can do for this? Please advise.
From a technical standpoint, there is no hard requirement that the version of Kafka you use is within the range documented. You're free to try using a newer version than what's documented (the Ops Agent will not forbid you from doing this), and it "might" work; it's just not officially supported or tested.
That being said, I tried running version 3.2.3 through the Ops Agent integration tests and they all passed, so there's a decent chance that 3.2.3 would work fine for you.
I want to migrate the database from Apache Cassandra to Amazon DynamoDB.
I am following this user guide
https://docs.aws.amazon.com/SchemaConversionTool/latest/userguide/agents.cassandra.html
When I try to create a clone data centre for extraction it throws
If you read through that document, you'll find that the conversion tool supports very old versions of Cassandra: 3.11.2, 3.1.1, 3.0, 2.1.20.
There will be a lot of configuration items in your cassandra.yaml that will not be compatible with the conversion tool including replica_filtering_protection since that property was not added until C* 3.0.22, 3.11.8 (CASSANDRA-15907).
You'll need to engage AWS Support to figure out what migration options are available to you. Cheers!
I want to partition my table by date. We are using CloudSQL instance for hosting our Mysql db. Does Cloud SQL support vertical partitioning? If yes, would it be similar to Mysql?
Cloud SQL for MySQL is basically a fully managed MySQL Community Edition databases by GCP. As mentioned in the comment, if the feature is not supported by MySQL itself, unfortunately you cannot expect that it will appear as a feature in Cloud SQL as well.
Here's the doc per each supported version which will inform you that vertical partitioning is not supported in MySQL 8.0, 5.7, and 5.6 (you can switch the doc version on their page):
https://dev.mysql.com/doc/refman/8.0/en/partitioning-overview.html
I have a large Google Cloud SQL (Second Gen) instance, and I would like to upgrade my database version from MySQL 5.6 to 5.7. But database version option is disabled on the edit instance form.
Why it is disabled, do i have to create a new instance then export and import existing database? My database is too large, and it will be a long downtime.
Per the Cloud SQL Migration docs, the only way to migrate versions is to export your data, and re-import into a new instance. The documentation mentions going from 5.5 to 5.6, but I would believe that going 5.6 to 5.7 would follow the same procedure.
Minor version upgrades to MySQL through Google Cloud now appears to be supported, though only through API calls:
https://cloud.google.com/sql/docs/mysql/upgrade-minor-db-version#gcloud
The pertinent section if using the gcloud CLI:
gcloud sql instances patch $INSTANCE_NAME --database-version=$DATABASE_VERSION
Substitute your instance name for the $INSTANCE_NAME variable, and your target database version in place of $DATABASE_VERSION.
There appear to be two REST APIs also available - see the documentation for details.
I am trying to develop server-side using loopback with database connector.
However, I am quite confused with installing loopback on AWS.
reference for installing loopback on AWS
This website mentioned that only loopback of version 2.0 could be installed.
Yet, when I browse through loopback website, https://strongloop.com/strongblog/how-to-setup-push-notifications-private-mbaas-amazon-aws-part-1/, this website shows that it seems possible to install loopback of version higher than 2.0 on AWS. Since there are some features only available after version 2.1x, it would be nice if AWS allows installation of loopback of version higher than 2.0. Could anyone help me solve the problem? BTW, I am only using free tier of AWS and do not intend to pay at this moment.
Even if you install the image that comes preconfigured with Loopback 2, you should be able to upgrade to newer versions using npm as you normally would (sudo npm install -g strongloop and the like). Imagine if there's a security issue that you'd need that wasn't backported for whatever reason...Loopback is just files and the image is just linux. You have free reign to update/upgrade whatever you need.
My recommendation would be to start out with a minimal Ubuntu image and install everything with npm. You'll understand the ecosystem better and won't be surprised by something you don't remember installing specifically.
One caveat that a bunch of preconfigured images have is they are only available on older instance types (m1 for instance--pun not intended ;)). They are slower and more expensive than newer instances.
You can probably install whatever you want.
The difference is that if you want a machine image that contains everything already there and don't want to build it yourself you're stuck to 2.0 here (https://aws.amazon.com/marketplace/pp/B00PG9I0M0)
What I would do is use the provided AMI and after that upgrade loopback manually.