I have a small virtual machine (f1-micro) on google cloud platform. It has been running daily without issues for more than a year. Some days ago I tried to install an sftp server. I installed some packages and create some users. Since then I'm not able to accces it :(.
Somebody knows what can I do to recover the access? Maybe, there is the option to move the data on the hard disk to another "fresh" virtual machine?
Thanks,
Ferran
Related
I've spent ages going around in circles on this so I'm hoping someone will point me in the right direction.
I'm creating a SQL Server lab running under Hyper-V on an Azure Windows Server 2016 Datacenter Gen 1 virtual machine. So far so good as I've got an AG running on two replica VMs, however I want to expand the lab to inlude a Windows Failover Cluster so I need to be able to create a shared disk and that's where I'm stuck. Whenever I try and add a shared disk in the Hyper-V Manager (or PowerShell) I get the following error:
The storage where the virtual hard disk is located does not support virtual hard disk sharing
It can't be the type of Azure disk I'm using as I get the same problem trying to create the Hyper-V shared disk on a Standard HDD, Standard SSD and Premium SSD with sharing enabled so what else do I need to do?
Regards,
Gordon.
You can try below possible solutions:
Please check if you have enabled the Sharing while creating the managed disk in the portal. If not then please enable it while deploying and add the max shares as per your requirement .
Reference:
Share an Azure managed disk across VMs - Azure Virtual Machines | Microsoft Docs
Enable shared disks for Azure managed disks - Azure Virtual Machines | Microsoft Docs
Please try to attach the VHDx file to IDE controller as its a Gen 1 virtual machine because only the SCSI controller has the option "virtual hard disk sharing" and Gen 1 VM only can boot from IDE.
Note: Please do not use this feature without CSV or Scale-Out File Server with SMB 3.0 on file-based storage .
Reference :
the storage where the virtual hard disk is located does not support virtual disk sharing (microsoft.com)
Thank you Prabhu Dutta Mohanty for providing the reference link .
Note: If the issue is still not resolved , Please create a support request to Azure support from portal (Support+Help) for assistance.
I have just installed VMware ESXi 7 as a virtual machine just for learning. I have seen it is feasible to create nested vms using VMware Player Workstation plus Intel chipset: my testing purpose is to create a virtual machine inside a virtualized ESXi server.
Actually I cannot install any vm, probably due to the fact I have not created any datastore yet.
In order to create a datastore I thought to edit the partition of the free space avalaible (for a linux vm 20GB are enough), but when I try to edit partition I get such summary in which I cannot configure anything at all (see pics).
Have you any suggestion?
When you install SO it's not a good practice to add it as a datastore. Please turn off you VM and add another disk to ESXI. After you boot up server again you will be able to create a new datastore.
I have a production server that hosts 3 VM's over Esxi 5.5.
Back in the day, I used a customized HP image to get ESXi installed on the Proliant server.
I have purchased a new server with Esxi 6.7 installed and wonder if I can move my 3 VM's hosted on the old HP server onto my new server (running ESXi 6.7).
The HP server sits 1500Km away so is challenging to test.
Did anyone come across any challenges removing VM's from one Host to another running different ESXi versions?
Thank you
You won't have any problems in the transportation process, but need to decide which transportation method use.
You can install vCenter Server Appliance then migrate with Storage vMotion.
If you do not want to install vCenter, you can turn the VMs power off and get an OVF copy to export (more on OVF here on vmware.com's site). You can then add this again from the deploy OVF section.
Link below for Vcenter installation.
https://www.tayfundeger.com/vcenter-server-appliance-6-7-kurulumu-bolum-1.html
https://www.tayfundeger.com/vcenter-server-appliance-6-7-kurulumu-bolum-2.html
Thanks.
VM objects are fairly backwards compatible and most go back quite a few years and a handful of versions, so you should be fine between those particular versions.
The biggest consideration is normally how to get the VM object data from point A to point B. Example:
Are you using storage based replication?
Are you SCPing the data directly from the hosts?
Are you exporting the VMs, transporting the data, and importing them?
Etc.
Yes, you can. The things you must take into account are:
1 - Hardware version. It's not possible to downgrade HW version through the ESXi UI. This won't be a problem in your case, as you are moving the VMs to a higher ESXi version that still supports ESXi 5.5's HW version. Once you have the VMs in the target server you can decide to upgrade to the most recent HW version for your new platform.
2 - VMFS version. ESXi 6.7 allows the use of VMFS-5 or VMFS-6, which is a newer version of the VMWare file system. You can indeed move VMs from VMFS-5 to VMFS-6. Nonetheless, unless it's unavoidable to do so, I would use the same VMFS version, as performing a cross-file system migration can make you fall into some incompatibilities that you should avoid.
3 - You will have to move your VMs over IP. If you don't own a VMWare license that allows you to migrate them, you can use an ESXi backup tool from 33hops.com that is compatible with unlicensed Free ESXi.
This XSIBackup-DC is a well-tested tool that allows to live migrate VMs over IP in licensed or unlicense versions of ESXi.
I'm looking for ways to migrate the server from physical to GCP cloud but there is a lot of challenges to be considered.
My plans are :
Lift and shift the data | thinking of this if not using velostrata
Migrate using GCP velostrata.
Migrate using velostrata was not so clear there is no defined way to do it. link -> https://cloud.google.com/migrate/compute-engine/docs/4.5/how-to/prepare-vms-servers/physical-servers
By going through the documentation it looks to be migrated to VMware first then to the GCP cloud.
Can you guys let me simplified the steps and confirmation on this?
GCP has a couple of options to migrate instances.
Import disk
The import tool supports most virtual disk file formats, including VMDK and VHD
This feature has the following limitations:
Linux virtual disks must use grub as the bootloader.
UEFI bootloaders are not supported for either Windows or Linux.
Linux virtual disks must meet the same requirements as custom images,
including support for Virtio-SCSI Storage Controller devices.
When installed on Windows virtual disks, application-whitelisting
software, such as Cb Protection by Carbon Black, can cause the import
process to fail. You might need to uninstall such software prior to
import.
If you are importing a virtual disk running RHEL, bring your own
license (BYOL) is supported only if the python-boto package is
installed on the virtual disk prior to import.
Operating systems on virtual disks must support ACPI
If you decide to go this route I recommend you to look and use the compatibility precheck tool
Velostrata
Velostrata supports 4 different sources of machines.
On-premise VM
Azure
AWS
Physical server
The guide you share indicates that you need to download "Migrate for Compute Engine Connector ISO image" (included in the link), save it in an USB and make it bootable.
Then you will need to continue with the steps here
You can also use the path you suggest to do a P2V migration to VMware environment using a tool such as VMconverter
Once your machine is in a VMware environment follow the on-premise Velostrata migration guide
We have the need to perform tests on localized platforms that put some burden on our hardware resources because for just a few weeks we might need plenty of servers and clients (Windows 2003 and Windows 2008, Vista, XP, Red Hat, etc) in multiple languages.
We typically have relied on blades with Windows 2003 and VMWare, but sometimes these are overgrown by punctual needs and also have the issue that the acquisition and deployment process is quite slow if the environment needs to grow.
Is Amazon EC2/S3 usable in the following scenario?
Install VMWare (Desktop because we need the ability to have snapshots) on an Amazon AMI.
Load existing VMWare images from S3 and run them on EC2 instances (perhaps 3 or 4 server or client OSes on each EC2 instance.
We are more interested in the ability to very easily start or stop VMware snaphsots for relatively short tests. This is just for testing configurations, not a production environment to actually serve a user workload. The only real user is the tester. These configurations might be required for just a few weeks and then turned off for a few months until the next release requires them again.
Is EC2/S3 a viable alternative for this type of testing purpose?
Do you actually need VMWare, or are you testing software that runs in the VMWare VMs? You might actually need VMWare if you are testing e.g. VMWare deployment policy, or are running code that tests the VMWare APIs. Examples of the latter might be you are testing an application server stack and currently using VMWare to test on many platforms.
If you actually need VMWare, I do not believe that you can install VMWare in EC2. Someone will correct & enlighten me if this is not the case.
If you don't actually need VMWare, you have more options. If you can use one of the zillion public AMIs as a baseline, clone the appropriate AMIs and customize them to suit your needs (save the customized version as a private AMI for your team). Then, you can use as many of them as you like. Perhaps you already have a bunch of VMWare images that you need to use in your testing. In that case, you can migrate your VMWare image to an EC2 AMI as described in various places in Google, for example:
http://thewebfellas.com/blog/2008/9/1/creating-an-new-ec2-ami-from-within-vmware-or-from-vmdk-files
(Apologies to the SO censors for not pasting the entire article here. It's pretty long.) But that's a shortcut; you can always use the documented AMI creation process to convert any machine (VMWare or not) to an AMI. Perform that process for each VMWare VM you have, and you'll be all set. Just keep in mind that when you create an AMI, you have to upload it to S3, and that will take a lot of time for large VMs.
This is a bit of a shameless plug, but we have a new startup that may deal with exactly your problem. Amazon EC2 is excellent for on-demand computing, but is really targeted at just a single user launching production servers. We've extended EC2 to make it a Virtual Lab Management environment, with self-service, policies and VM sharing. You can check it out at http://LabSlice.com and see if it meets your needs.
Amazon provides a solution themselves now: http://aws.typepad.com/aws/2010/12/amazon-vm-import-bring-your-vmware-images-to-the-cloud.html