It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 10 years ago.
Will it be installed on existing OS or hardware?
What's the difference between Hypervisor and virtualbox? I know I can install virtualbox on my windows xp and then install a couple of VMS with linux or windows on it.
thanks.
Where is Xen Hypervisor installed? Will it be installed on existing OS
or hardware?
As explained in the fine wikipedia article, Xen is a bare-metal hypervisor, meaning it runs directly on the hardware.
However, as Xen itself has no user interface, it always requires at least one installed guest OS to work. There must be one OS guest that runs with special privileges (the so-called "dom0"). This dom0 is used to manage Xen, and so has a similar role to the host OS on a hosted hypervisor. So, while technically speaking Xen runs directly on your hardware, in practice you will still install an OS, then install Xen from inside the OS, just like for a hosted solution. However, you will then have to reboot into Xen (which will boot the dom0 OS).
What's the difference between Hypervisor and virtualbox?
Hypervisor is the general term, and both Xen and VirtualBox are implementations of hypervisors.
I know I can install virtualbox on my windows xp and then install a couple of VMS
with linux or windows on it.
Yes, you can do the same with Xen (though support for Windows guests seems to be more limited than with VirtualBox).
Full virtualisation is difficult to achieve (e.g.: virtualbox) and is slow, because you should create an exact copy of the underlying hardware, and difficult to recover from hardware errors.
Paravirtualisation (xen hypervisor) provides the machine hardware, but sitll manages it. And Operating System, that will be installed as VM should have some modifications.
Xen still needs some OS to manage resources of the VMs, so it will not be installed on hardware only. Citrix Zen Server for example uses CentOS5 (which is quite outdated, by the way).
The paravirtualisation approach is good for Cloud computing because of the ability to effectively run multiple VMs on a number of different machines and scale.
I suggest you to read an article called "Above the Clouds".
Related
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 8 months ago.
Improve this question
I have a use case that I want to install windows 10 on an aws instance. Then on top of it, I want to install VMware workstation. In that VMware workstation, i want to install multiple VMs e.g kali, redhat, etc. Earlier this week, i had a simple aws instance( with server 2016) and it didn't allowed me to install VMs on vmware workstation inside server2016. It said that hypervisor and VMware can't stand simultanously. While looking for the resolution, I found exact same issue like mine:
https://forums.aws.amazon.com/thread.jspa?threadID=293113
And it said some thing like this:
Nested virtualization is not supported on AWS instances unless you are using AWS bare metal instances. https://aws.amazon.com/blogs/aws/new-amazon-ec2-bare-metal-instances-with-direct-access-to-hardware/
Now please clearly tell me that "if i get c5.xlarge bare metal instance of aws, then can I install my use case as i described in my first paragraph?" Please help. I couldn't find exact answer anywhere else!
Thank you in advance...
There is no such thing as a c5.xlarge bare metal instance.
Instances run on a physical 'host' in the AWS data center. Each host supports one 'family' of instances, such as C5. This is because each family has a specific type of processor and a particular ratio between CPU and RAM.
A C5 host has 96 vCPUs and 192 GB of RAM. This can be divided into different 'instance types' within the family, such as:
c5.large with 2 vCPUs and 4 GB RAM
c5.xlarge with twice as much (4, 8)
c5.12xlarge with 12 times as much as a c5.xlarge
All the way up to c5.24xlarge that has all 96 vCPUs and 192 GB of RAM
The instance type you choose basically gives you a 'slice' of the host.
If you wish to go bare metal, then you get the entire host with 96 vCPUs and 192 GB of RAM. When selecting bare metal, you get the whole host computer and it is big!
This is why you cannot get a c5.xlarge as a bare metal instance.
So, your choices are:
Get a c5.metal instance, install VMWare and create smaller virtual computers, or
Use VMware Cloud on AWS where VMware runs the system for you and you can get smaller virtual computers, or
Give your students Amazon EC2 instances (which would be the simplest option!), or
Run your own hardware
I think azure cloud are supporting nested virtualization.
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
A task assigned to me is to virtualize the Lab using VMware vSphere products (ESXi and vSphere Client). I spent couple of days to get know-how about vSphere and finally decided to put my hands on it......While installation begins I proceeded very carefully and i reached this Warning message from installer:
Warning: This disk will be repartitioned.
What I understand from this warning message that my all hard disk is going to be repartitioned and obviously formatted and em gona loose my all data.
I was about to confirm but suddenly realize that "repartition" mean my all data on the disk will be gone, and this is what I really don’t like to face.
My problem is that i want to install VMware ESXi 5.1 on a machine which has single hard disk with 4 partitions(3 primary, 1 extended). All three primary partitions are hosting one OS; mean 3 Operating Systems are already installed on each primary partition.
So I want to install VMware ESXi on one of the primary partition, i am ready to lose OS on that primary partition but I do not want to lose my other 2 OS and data on extended partition.
How can I install VMware ESXi 5.1 alongside my other OS without losing data???
is it possible??? If yes guide me please guide, if not then give me kind suggestions.
i come here after googling. ;-(
Bundle of thanks..........
Unfortunately, if you only have one local disk, the ESX installation will reformat and repartition that drive. There is some good news though!
Instead, you can download ESXi, which is a free hypervisor and is considered "installable". So, you can put the ESXi software onto something like an external HD or a USB stick and you will be able to boot your ESXi server using whichever external drive you select. This will also allow you to keep the partitions on your local disk in place while you use that disk's remaining space for your VM datastores.
Here is a great link from VMware that shows you exactly how to do it, enjoy!
http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1020655
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 9 years ago.
Improve this question
Can anyone give me a simple comparison of those two? It is hard to get the idea from their web site.
VM Player runs a virtual instance, but can't create the vm. [Edit: Now it can.] Workstation allows for the creation and administration of virtual machines. If you have a second machine, you can create the vm on one and run it with the player the other machine. I bought Workstation and I use it setup testing vms that the player runs. Hope this explains it for you.
Edit: According to the FAQ:
VMware Workstation is much more advanced and comes with powerful features including snapshots, cloning, remote connections to vSphere, sharing VMs, advanced Virtual Machines settings and much more. Workstation is designed to be used by technical professionals such as developers, quality assurance engineers, systems engineers, IT administrators, technical support representatives, trainers, etc.
VMWare Player can be seen as a free, closed-source competitor to Virtualbox.
Initially VMWare Player (up to version 2.5) was intended to operate on fixed virtual operating systems (e.g. play back pre-created virtual disks).
Many advanced features such as vsphere are probably not required by most users, and VMWare Player will provide the same core technologies and 3D acceleration as the ESX Workstation solution.
From my experience VMWare Player 5 is faster than Virtualbox 4.2 RC3 and has better SMP performance. Both are great however, each with its own unique advantages. Both are somewhat lacking in 2D rendering performance.
See the official FAQ, and a feature comparison table.
from http://www.vmware.com/products/player/faqs.html:
How does VMware Player compare to VMware Workstation? VMware Player
enables you to quickly and easily create and run virtual machines.
However, VMware Player lacks many powerful features, remote
connections to vSphere, drag and drop upload to vSphere, multiple
Snapshots and Clones, and much more.
Not being able to revert snapshots it's a big no for me.
One main reason we went with Workstation over Player at my job is because we need to run VMs that use a physical disk as their hard drive instead of a virtual disk. Workstation supports using physical disks while Player does not.
Workstation has some features that Player lacks, such as teams (groups of VMs connected by private LAN segments) and multi-level snapshot trees. It's aimed at power users and developers; they even have some hooks for using a debugger on the host to debug code in the VM (including kernel-level stuff). The core technology is the same, though.
re: VMware Workstation support for physical disks vs virtual disks.
I run Player with the VM Disk files on their own dedicated fast hard drive, independent from the OS hard drive.
This allows both the OS and Player to simultaneously independently read/write to their own drives, the performance difference is noticeable, and a second WD Black or Raptor or SSD is cheap.
Placing the VM disk file on a second drive also works with Microsoft Virtual PC.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 3 years ago.
Improve this question
I currently work in an organization that forces all software development to be done inside a VM. This is for a variety of risk/governance/security/compliance reasons.
The standard setup is something like:
VMWare image given to devs with tools installed
VM is customized to suit project/stream needs
VM sits in a network & domain that is isolated from the live/production network
SCM connectivity is only possible through dev/test network
Email and office tools need to be on live network so this means having two separate desktops going at once
Heavyweight dev tools in use on VMs so they are very resource hungry
Some problems that people complain about are:
Development environment runs slower than normal (host OS is windows XP so memory is limited)
Switching between DEV machine and Email/Office machine is a pain, simple things like cut and paste are made harder. This is less efficient from a usability perspective.
Mouse in particular doesn't seem to work properly using VMWare player or RDP.
Need a separate login to Dev/Test network/domain
Has anyone seen or worked in other (hopefully better) setups to this that have similar constraints (as mentioned at the top)?
In particular are there viable options that would remove the need for running stuff in a VM altogether?
In particular are there viable options
that would remove the need for running
stuff in a VM altogether?
Given that you said there are unspecified risk/governance/security/compliance reasons for your organization's use of VMs, I doubt any option we could provide could negate those. Ultimately it sounds like they just need their development team as sandboxed as possible.
(And even so, the question/answers would probably be better off at serverfault since it's more networking/security oriented.)
It sounds like a big problem is not having enough horsepower on the host OS. WinXP should be fine, but you need to have adequate hardware. i.e. at least 3 GB RAM, dual core CPU, and hardware that supports virtualization. Clipboard sync should be working with the VM.
I am not currently doing this, but I've thought about it, and we're kind of kicking this idea around with the idea of making it easier to standardize the dev environment, and to avoid wasting a day when you get a new PC. I'm dismayed to hear that it's not the utopia that I had dreamed...
I've been using VMs as a development environment for a long time. There's nothing inherently wrong with it, and it presents lots of benefits.
Ensuring a consistent environment
Separating file systems for different backup scenarios
Added security
Potentially gives developers access to more raw computing power.
There is a lot of innovation in the VM world, as evidenced by the growing popularity of VM farms, hardware support for virtualization, and controlled "turnkey" solutions, like MS's VirtualPC images for testing browser compatibility and the TurnKey set of appliances.
As others have said, your issues are probably due to insufficient hardware or sub-optimal configurations.
Development environment runs slower than normal (host OS is windows XP so memory is limited)
This should not be noticeable. XP vs. Windows Vista or Win7 is a marginal comparison. I would check the amount of physical RAM allocated to the VM.
Switching between DEV machine and Email/Office machine is a pain, simple things like cut and paste are made harder. This is less efficient from a usability perspective.
There are VM-specific optimizations/configurations that can make these tasks seamless. I would consult your VM maintenance staff.
Mouse in particular doesn't seem to work properly using VMWare player or RDP.
Again, should be seamless, but consult VM staff.
Need a separate login to Dev/Test network/domain
I would see this as a business decision: your company could obviously set up virtual machines with the same domain poicies as your own personal workstation, but may have other (big brother?) purposes for forcing you to login separately.
As far as using VM's as an agent of control, I think there are better solutions, like well-designed authorization controls around the production machines. There's nothing like paper trails to make people behave themselves.
I wondered if anyone uses virtualized desktop PCs (running WinXP Pro or older) to have some old applications that are seldom used available for some ongoing tasks.
Say you have a really old project that every once in a while needs a document update in a database system or something like that. The database application is running on a virtualized desktop that is only started when needed.
I think we could save energy, hardware and space if we would virtualize some of those old boxes. Any setups in your company?
edit Licensing could be of concern, but I guess you have a valid license for the old desktop box. Maybe the license isn't valid in a VM environment, I'd definitly check that before.
Sure enough, if the application is performance critic, virtualization could hurt. But I'm thinking about some kind of outdated application that is still used to perform, say a calculation every 12 weeks for a certain customer/service.
I use virtualized desktops for:
Support that requires VPN software I do not want on my own desktop. This also lets a whole team share the support computer for a specific customer.
A legacy system which we use several different versions of (depending on customer's version) and they're not really compatible so its good to have a virtualized desktop for each version.
We use virtualisation to test on a variety of Operating Systems - the server application runs under linux, and we have a production (real) server, and a couple of test servers, which are all VMs.
The client runs under Windows, which, being an OS X user I have to run in a VM, and the other developer I work with runs an XP VM on his 8-core Vista box.
(I also have a seperate VM for running CAD software, but that's not really programming)
It depends on the requirements of the legacy systems. Very often if a system is relient on a certain clock frequency, then it better and morereliable to keep the older OS systems running as Virtulized OS' can do funy things to performance.
If the legacy systems aren't that critical, then go for it! One piece of advice I would give is to ensure that the system works FULLY before chucking out your old 3.11 systems as I have been stung before! To fully perform the testing can cost more money then you might save, but its up to anyone who make the decisions to ensure that is considered.
We use virtualisation for testing out applications on Vista. Or rather customers do the testing and we use virtualisation to reproduce the bugs they complain about.
I guess the thing that would stop me from using lots of virtual instances of my favourite proprietary OS would be licencing. I presume Microsoft would want me to have a licence for every installation, virtual or otherwise?
We use VMWare with a virtual windows XP here at work to run some old development tools with very expensive licenses that don't run at all on Vista. So VMWare saved us about $5000 in licenses.
Since my last machine upgrade I have been running virtualised OS's for a number of tasks. For example I use a different set of Visual Studio plugins for managed and c++ unmanaged development. Some things I found:
Run your vmware setup on a machine with plenty of resources. I'll repeat...plenty of resources! A fast quad and 8GB of memory is what my current machine is running and it runs sweet (warning you need a 64bit OS for the 8GB!).
I wouldn't worry about app performance if your current physical hardware is old (2+ years). With a decent machine I find the virtualized apps run faster than on the legacy hardware!
When upgrading to a new workstation, p2v your old workstation. No need to worry about synergy or a KVM in the transition period any more!
I've used virtualisation so I could take my development environment around with me while travelling. As long as I could install MS Virtual PC, (and the PC/laptop had generous enough RAM) then I could access all my tools, VPN, Remote desktop links, SQL databases etc...
Worked fairly well, just a little slower than I like. I could have carted a laptop around, but found a small portable harddrive to be lighter/easier and just as effective.
However, consulting for several clients - all with different VPN requirements/passwords/databases/versions of frameworks & tools etc, I've found that having a Virtualised support environment for each is well worth it. Then multiple users have access to what is needed when supporting each client - they just need to either remote desktop (or run directly) the virtualised instance.
I've used VMs to handle work-related tasks that I didn't want / couldn't do on the company-issued laptop. Specifically, I needed to have several editions of the JRE running at the same time, which Java doesn't really like.
To get around this, I built several VMs that each ran the one tool I needed in trimmed-down XP instances.
Another thing to consider is that if you have a 5-yr-old server running some app, it's probably going to run just fine on a VM on new hardware. So, if you have a rack of old devices, buying one or two "real" servers, installing something like ESX (I'm most familiar with that tool, though Xen and others exist), then use a physical-to-virtual conversion tool to get those old devices switched to VMs so you can reduce your electricity consumption, management headaches, and worries about a critical device failing and not being able to find hardware for it.
We use VM for legacy apps, and have retired old machines that served up those apps. It eliminated the concern of matching drivers from NT to Win2k3. From a disaster recovery perspective this also helped as we couldn't find boxes to support the old apps at the DR data center.
The likes of VMWare are invaluable tools for browser testing of web applications. You can pretty easily test many combinations of OS and browser without having rank upon rank of physical machines running that software.