Does GCP ops agent receiver support kafka version 3.2.x? - google-cloud-platform

I am trying to collect logs/metrics from kafka connector in GCP.
I read the documentation
It seems that ops agent only support till Apache Kafka Version 3.0.0 and I am using Kafka version 3.2.3.
May I know if there is any roadmap of version support of Kafka in ops agent?
Or any alternatives I can do for this? Please advise.

From a technical standpoint, there is no hard requirement that the version of Kafka you use is within the range documented. You're free to try using a newer version than what's documented (the Ops Agent will not forbid you from doing this), and it "might" work; it's just not officially supported or tested.
That being said, I tried running version 3.2.3 through the Ops Agent integration tests and they all passed, so there's a decent chance that 3.2.3 would work fine for you.

Related

AWS DocumentDb not support mongodb 4.0

I do not know why AWS DocumentDb does not support MongoDB version that above 3.6? Should I use mongo 3.6 or 4.0 above?
Amazon DocumentDB now supports MongoDB 4.0 compatibility including transactions: https://aws.amazon.com/about-aws/whats-new/2020/11/amazon-documentdb-with-mongodb-compatibility-adds-support-for-mongodb-4-and-transactions/
Document DB is compatible with only MongoDB 3.6.
See: https://aws.amazon.com/documentdb/features/
Whether you want to use 3.6 vs 4.0 or even 4.2 or 4.4 is very subjective to what you want to do with the DB.
The pros of using Document DB is that its a managed service and hence you dont have to worry a lot about setting it up.
The con is you will not get features present in version 3.6 onwards example multi document transactions, new operators in the agg pipeline, bug fixes etc.
To figure out the exact changes check - https://docs.mongodb.com/manual/release-notes/
You can install mongodb on EC2 instances, this will ensure you will get the latest and greatest mongodb. However this comes with the added work of managing the mongodb instance, its backups, High availability considerations etc.
Do note: No matter what you decide I would recommend that you try to use the latest drivers that are present today so that you have the freedom to go to the latest version of self installed mongo or even upgrade document db engine versions as they become available.
AWS DocumentDb uses its own database engine compatible with MongoDB 3.6 API.
MongoDB 4.0 API is not supported as of yet.
We have MongoDB Atlas available in AWS - AWS also supports fully managed database service - MongoDB Atlas - https://aws.amazon.com/de/quickstart/architecture/mongodb/
AWS enables to set up the infrastructure to support MongoDB deployment in a flexible, scalable, and cost-effective manner on the AWS Cloud.
https://aws.amazon.com/de/quickstart/architecture/mongodb/

How often do Cloud Build Node.js versions update?

I couldn't stomach purchasing the $150 for GCP's support service for this one question. I'm just looking to understand the schedule for Cloud Build Node.js versions. It's still stuck on Node.js v10.10 and my projects are starting to require higher versions to build. According to Cloud Build's changelog, I don't believe the Node.js version has updated in years. Any ideas?
As per the official Github repository:
Our client libraries follow the Node.js release schedule. Libraries are compatible with all current active and maintenance versions of Node.js.
So, this means it should work with Node.js 12 and the updates should be more constant. In addition to that, in here, it says that if you are using a Cloud Build config file, you can use Node.js 12, so the Node.js' latest version should be compatible with Cloud Build.
To summarize, by the repository, it should follow Node.js schedule. However, in case you think this is not occurring, I would recommend you to raise a bug on the Google's Issue Tracker - it's free, by the way - so they can assess this.

Is credhub available with run.pivotal.io?

If yes then what is the url/api to access it from run.pivotal.io 's trial account.
Also can the features of credhub be tested using pcfdev?
Thanks
Ver 0.28.0 is the latest PCFDev available on PivNet
It is PCF ver 1.11.x. Two versions older. This version does not support CredHub.
If I recall correctly, in ver 1.12 all internal properties are stored in CredHub. That said, PCFDev is not meant for production by any means. So, it may not support CredHub.
To test locally, you can spin up a credhub instance. And use that in your PCF app. To do that you will need VirtualBox and spin up a bosh-lite instance with CredHub.
Take a look at this article on how to set it up.
No, credhub is not supported for application developers on PWS at this time.

Can I use loopback of version higher than 2.0 on AWS?

I am trying to develop server-side using loopback with database connector.
However, I am quite confused with installing loopback on AWS.
reference for installing loopback on AWS
This website mentioned that only loopback of version 2.0 could be installed.
Yet, when I browse through loopback website, https://strongloop.com/strongblog/how-to-setup-push-notifications-private-mbaas-amazon-aws-part-1/, this website shows that it seems possible to install loopback of version higher than 2.0 on AWS. Since there are some features only available after version 2.1x, it would be nice if AWS allows installation of loopback of version higher than 2.0. Could anyone help me solve the problem? BTW, I am only using free tier of AWS and do not intend to pay at this moment.
Even if you install the image that comes preconfigured with Loopback 2, you should be able to upgrade to newer versions using npm as you normally would (sudo npm install -g strongloop and the like). Imagine if there's a security issue that you'd need that wasn't backported for whatever reason...Loopback is just files and the image is just linux. You have free reign to update/upgrade whatever you need.
My recommendation would be to start out with a minimal Ubuntu image and install everything with npm. You'll understand the ecosystem better and won't be surprised by something you don't remember installing specifically.
One caveat that a bunch of preconfigured images have is they are only available on older instance types (m1 for instance--pun not intended ;)). They are slower and more expensive than newer instances.
You can probably install whatever you want.
The difference is that if you want a machine image that contains everything already there and don't want to build it yourself you're stuck to 2.0 here (https://aws.amazon.com/marketplace/pp/B00PG9I0M0)
What I would do is use the provided AMI and after that upgrade loopback manually.

Node.js 0.8.2 on CloudFoundry Micro edition 1.2

The documentation for CloudFoundry indicates that Node.js version 0.8.2 is supported. However, when I run "vmc runtimes" against my local micro edition, it only lists node and node06 (not node08)
I'm running 0.8.2 locally on my box -- how can I upgrade Micro edition so that it also has 0.8.2 installed?
Micro Cloud Foundry 1.2 was released before Node.js 0.8.2 support was added. MCF currently cannot be upgraded to support it. However, a new MCF version is planned for the very near future which will bring parity to all the feature on CloudFoundry.com. We also plan a faster release cadence of MCF so it doesn't fall behind on features.
It's not completely trivial to patch MCF 1.2 to add node 0.8, you'd need to install the relevant packages as well as add information to the Cloud Foundry manifest files to enable it to advertise the node08 runtime.
However, the next Micro Cloud Foundry release should be available pretty soon so if you can hold on just a little longer, we should have some good news for you :-)
Out of interest, what application are you looking to build / run that requires 0.8.x?
Micro Cloud Foundry has just been updated to be in sync with the codebase on github so it now supports Node.js 0.8.2. Read the announcement here: http://blog.cloudfoundry.com/2012/11/08/new-release-of-micro-cloud-foundry/