Create a Private Environment with Power Platform - microsoft-dynamics

Hi I'm trying to setup my own environment in Power Platform.
A do have a community Environment licence, but i cannot complete all the Microsoft Learning workshops as this environment is limited in functionality.
I tried to create a practice environment at work. However, I do not have the "Developer" option when i try to create my own environment. I do have the "Sandbox" option as one of my types.
But these environments are by default to anyone, unless you explicit limit them to an AD group. I do not belong to an AD group exclusively. Nor can i create one just for myself.
Is there a way to create an Environment that is Private. Where only i can access it?
Erick

As you mentioned the Community Plan have some limitations... And in my understanding when you use a Community Plan you get your own environment where you are the only one with access.
I share with you this documentation link where explain more deeply the details of this Power Apps Plan.
https://learn.microsoft.com/en-us/powerapps/maker/dev-community-plan
Based on the screenshot taken from this link I understand that you are already using a Developer environment.

Related

Is it possible possible to create a shared folder between users in AWS Sagemaker Studio?

I'm currently trying to migrate a data science environment (jupyter notebook)running on Kubernetes to Sagemaker Studio.
I set up SSO and I now have privates work spaces for each user but I'd like to also have a shared folder between all the users. I've googled quite a bit to find an answer to this question without success.
Thanks for you help
As of today, there isn't an option to have a shared folder between users in Studio. Each user profile gets a dedicated directory on EFS that's private to them.
If you're sharing data, S3 is an alternative option.
I work at AWS and my opinions are my own.

Data Science/Engineering (Dev/Prod) Environment

I am going to create environments. For now i have gcp machine and i run jupyter in there. Everytime, i need start it, and with 3 people it is hard to work in same environment. I know, there is docker, jupyter hub, but did not find and suitable roadmap to create dev/prod environment.
My aim to create dev and production environment. Everything should be on GCP.
Any suggested path ?
Thanks
You can take a look at the best practices for enterprise organizations. In order to properly split resources it's often advised to use different projects. However, depending on the GCP product, you could also use versions, such as with App Engine (see this StackOverflow thread).

Would like to get build information from Google Cloud Profiler

I'm using Google Cloud Profiler (located at https://console.cloud.google.com/profiler) and would like to know how my profiling data changes across different builds of my application.
One way to do that would be to check the range of dates during which a particular commit was running on production, but that's time consuming because I have to:
Get the start date/time of release, determine the date/time of the next release
Set those dates manually in the profiler interface from the link above
That's really not terrible, but it'd be great to be able to set BUILD_ID environment variable like I can in Cloud Build and then be able to access that from the UI. Is something like this possible? Or is my approach the best way to do this at the moment?
Comparing across service versions would likely be a simpler and more precise way to do this (as opposed to using the time interval to select for profiles). To compare across service versions, it is necessary that the profiling agents set the service version.
The service version can be specified in the configuration passed to the agent (for the Go, Python, or Node.js agent) or via the -cprof_service_version flag (for the Java agent). If one is setting the service version using the configuration passed to the agent (applicable for the Go, Python, and Node.js agents), it may be convenient to use a flag or command line argument to set the service version so that the source code won't need to updated with each new version.
If one is running on Knative or App Engine standard, the service version should be auto-populated. These environments set the K_REVISION and GAE_VERSION environment variables (respectively), and the profiling agents (for all supported languages) use these environment variables to populate the service version. If one is running in another environment and modifying the source code is inconvenient or infeasible, one can set either the K_REVISION or GAE_VERSION environment variable in the environment running the application with the agent enabled to specify the service version.
My understanding is that the BUILD_ID is available at build time, but not at run time, so I don't know that it's possible for agents to use that directly.
(Disclosure: I work on Cloud Profiler at Google)
You can set the service version for this purpose. Please refer to the agent documentation for how to set it for supported languages.
For example, this shows using ServiceVersion for Go services.

Create gMSA with managed AD on google cloud

For a POC for using AD on the google cloud with kuberenetes, I created a managed active directory, as is described in this link.
To add a gMSA account for the AD, I looked at this documentation.
It looks like I should use the New-ADServiceAccount command from the AD VM.
However, when looking at the domain I've created on pantheon , I couldn't find the VM it is on.
The interface does not give me any clues as to how I could add users, or do anything with the domain.
Any help will be appreciated.
Thank you,
It not written in the docs, but it is not complicated.
One should first join the domain from another windows machine (note that this is not supported on all windows OSes). This is done exactly like a regular machine joining a regular AD, as long as the domain and the machine are in the same vpc.
Then, install the AD roles on the joined machine, and run the New-ADServiceAccount commnad from there.
Comment 1:
Some machines on the GCE cannot join a domain (even with windows server datacenter). I have not figured out why, and instead opened a different machine.
Comment 2:
Windows integration in kubernetes is new, and possibly this answer will not be relevant in 6 mounts.

GCP Deployment Manager - What Dev Ops Tool To Use In Conjunction?

I'm presently looking into GCP's Deployment Manager to deploy new projects, VMs and Cloud Storage buckets.
We need a web front end that authenticated users can connect to in order to deploy the required infrastructure, though I'm not sure what Dev Ops tools are recommended to work with this system. We have an instance of Jenkins and Octopus Deploy, though I see on Google's Configuration Management page (https://cloud.google.com/solutions/configuration-management) they suggest other tools like Ansible, Chef, Puppet and Saltstack.
I'm supposing that through one of these I can update something simple like a name variable in the config.yaml file and deploy a project.
Could I also ensure a chosen name for a project, VM or Cloud Storage bucket fits with a specific naming convention with one of these systems?
Which system do others use and why?
I use Deployment Manager, as all 3rd party tools are reliant upon the presence of GCP APIs, as well as trusting that those APIs are in line with the actual functionality of the underlying GCP tech.
GCP is decidedly behind the curve on API development, which means that even if you wanted to use TF or whatever, at some point you're going to be stuck inside the SDK, anyway. So that's why I went with Deployment Manager, as much as I wanted to have my whole infra/app deployment use other tools that I was more comfortable with.
To specifically answer your question about validating naming schema, what you would probably want to do is write a wrapper script that uses the gcloud deployment-manager subcommand. Do your validation in the wrapper script, then run the gcloud deployment-manager stuff.
Word of warning about Deployment Manager: it makes troubleshooting very difficult. Very often it will obscure the error that can help you actually establish the root cause of a problem. I can't tell you how many times somebody in my office has shouted "UGGH! Shut UP with your Error 400!" I hope that Google takes note from my pointed survey feedback and refactors DM to pass the original error through.
Anyway, hope this helps. GCP has come a long way, but they've still got work to do.