DigitalOcean - How to get the droplet cpu utilization via API - digital-ocean

Is there an api to retrieve the droplet's cpu utilisation (not the load metric)? I cannot figure out how to retrieve that information

Related

AWS LightSail RDS - How Much RAM Do I Need

I'm just setting up a hight availability WordPress network and I need to decide how much RAM I need for the database instance. On a web server you run "top" and find out how much RAM is being used per MySql process and then look in your config file and look at the maximum number of processes that are allowed to run.
How do you calculate how much RAM you will need on a High availability MySQL database running in AWS LightSail? The plans seem very light on RAM. For example a $20 webserver gets 4GB of RAM whereas a $60 database server get 1GB of RAM. Why is this and how many processes will 1GB run?

Google monitor - Dashboard with CPU, Memory and Disk usages with GCE filter

I'm trying to set up Googlecloud monitor Dashboard for my GCE's. I'm expiriancing some difficulties though when tring to filter.
I have serveral GCE, and some are not running and are as backup, but are still displayed in the Cloud Monitor.
I would like to monitor 3 metrics (for now) : CPU, Mem, Disk usage.
CPU wasnt a problem as i could just filter by GCE instance name:
But now if i try to do the same for Memory and Disk usage, I dont have the option to filter as I did using CPU. I tried serveral different approaches like filter by "metadata labels:name", "label", "zone" etc. - all result in a "no data avalible for selected timeframe" (without the filter data is displayed). I feel like I'm missing something trivial:
What am I doing wrong? How can I filter by Instancename? Do i need to Activate some logger on Google cloud? Thank you verymuch in advance!
Use the Cloud Monitoring agent to gather system and application metrics (disk, CPU, network and process) from VM instances and send them to Monitoring.
Install the Monitoring Agent
Use the Cloud Logging agent to gather logging metrics from VM instances and send them to Cloud Monitoring.
Install the Logging Agent

how to distribute surplus load of user traffic to google app engine from google compute VM ? running django with apache

I am running django on google VM instance using apache and mod wsgi... i however am unsure of the concurrent requests that my app shall receive from the users and would like to know if i can transfer the surplus load of the VM to the App engine automatically to prevent the server from crashing.
I am unable to find any solution expect running kubernetes cluster or docket containers to effectively manage the load. but in need to be free of this hassle and send off the excess load to GAE.
If you want to analyze the traffic, latency and load of your resources and applications, I would recommend you to start with Stackdriver Trace.
As per documentation, Stackdriver Trace is a distributed tracing system that collects latency data from your applications and displays it in the Google Cloud Platform Console. You can track how requests propagate through your application and receive detailed near real-time performance insights. Stackdriver Trace automatically analyzes all of your application's traces to generate in-depth latency reports to surface performance degradations, and can capture traces from all of your VMs, containers, or Google App Engine projects.
Once you have determine the user traffic or you have a better idea about this, then you can try using "Instance Groups".
GCE offers two kind of VM instance groups:
Managed instance groups (MIGs) allow you to operate applications on multiple identical VMs. You can make your workloads scalable and highly available by taking advantage of automated MIG services, including: autoscaling, autohealing, regional (multi-zone) deployment, and auto-updating.
Unmanaged instance groups allow you to load balance across a fleet of VMs that you manage yourself.

Idle CPU utilization on Google Compute Engine

Is there a reason why Google Developer Console shows 4-5% CPU utilization for my idle Debian Linux VM even though the VM itself claims it is 100% idle?
The CPU usage in Google Cloud Console is not that of the instance, but the CPU usage of the container managing it. This container is in charge of providing the virtualization services to the instance and collecting all the metrics. So, the Google Cloud Console CPU utilization shows the aggregate CPU usage for both the container and the instance.

Computing power of AWS Elastic Beanstalk instances

I have a CPU-intensive application that I'm considering hosting on 1+ AWS Elastic Beanstalk instances. If at all possible, I'd like to throttle it so that I don't dip over the "free" utilization of the instances.
So I need to figure out what kind of hardware/virtualized hardware the Beanstalk instances are running on, and compare that to the maximum CPU utilization of the free versions.
So for instance, if each Beanstalk instance is running on, say, 2GHz CPUs, and my app performs a specific "supercalc" operation that takes 50 million CPU operations, but the free version of the app only allows me to utilize 100 billion operations per day, then I am limited to 100billion/50million = 2,000 "supercalcs" per day on a free instance. So if the CPU is 2GHz, then my app instance could only run for 2GHz/50million = 40 seconds before I've already "maxed out" the free CPU utilization on the Beanstalk instance.
This is probably not a great example, but hopefully illustrates what I'm trying to achieve. I need to figure out how much I need to throttle my app, or how long my app could run before I max out the Beanstalk CPU utilization, and it really comes down to how beefy the AWS Beanstalk machines are. Thanks in advance!
Amazon EC2 instances aren't based on a "CPU utilisation" billing system (I think Google App Engine is?) - EC2 instance billing is based on the amount of time the machine is "on" regardless of what is doing. See the Amazon EC2 Pricing for the amount it costs to run the different instances sizes in different regions.
There is a special case which is the "Micro" instance - this provides the ability to have short bursts of higher CPU usage than the "small" instance at a lower cost, but if you overuse it you get throttled back for a period (which you don't with a Small). This isn't the same as having an operation limit though, and the price remains the same whether you're throttled or not.
Also note that with Elastic Beanstalk you're also paying for the Elastic Loadbalancer, any storage and bandwidth, and also any database service you are using.
Given all that though - AWS does have a Free Tier - however this is only for the first 12 months of a new account. The Free Tier will cover the cost of a micro EC2 instance, Elastic Loadbalancer, RDS database and other ancillary services - see the link for more info.