Sagemaker Studio worked flawlessly for me for the first 6 months. Then I started observing this issue. Screenshot of the error message
The screen holds up at this stage forever. Here's what I have tried:
Clearing my cache, even using a different machine. So I don't think the issue lies with the browser or my machine.
Pressing 'Clear workspace' in the screenshot above.
Shutting down all the apps in my sagemaker domain (excluding the 'default' app). This used to work initially but now this has stopped working all-together.
Created a new sagemaker domain with fraction of the files in the previous domain. Still, I see the same error message in the new domain as well.
This is severely affecting my work and I can't find a solution for this anywhere on the internet.
I've seen this issue before. Restarting the JupyterServer app and clearing out the EFS storage attached to your Studio domain helps.
Also, it is worth checking the JupyterServer logs to see what is causing the issue. Additionally, please make sure to use the latest version of Studio.
Related
I'm experiencing an error in Google Cloud Platform.
I'm stying to start my free trial in GCP, following the step I've:
Accepted Terms/Service conditions (Step 1)
Verified my account through phone (Step 2)
And here the console start loading (before showing the step 3) without finishing. I just waited for 12h, so I think there is a problem.
For some other info: I'm new on the platform; I used some service like Firebase but it's the first time I go for the GCP full package.
I know this is super late, but I've just run into the same issue and strangely enough the solution for me was to upgrade on my phone rather than on my desktop
I faced the same issue. Then after clearing the cache and disabling the ad-blocker and popup blocker, It worked. Give a try by disabling all the plugin installed in your web browser.
I am having some of my GCP instances behave in a way similar to what is described in the below link:
Google Cloud VM Files Deleted after Restart
The session gets disconnected after a small duration of inactivity at times. On reconnecting, the machine is as if it is freshly installed. (Not on restarts as in the above link). All the files are gone.
As you can see in the attachment, it is creating the profile directory fresh when the session is reconnected. Also, none of the installations I have made are there. Everything is lost including the root installations. Fortunately, I have been logging all my commands and file set ups manually on my client. So, nothing is lost, but I would like to know what is happening and resolve this for good.
This has now happened a few times.
A point to note is that if I get a clean exit, like if I properly logout or exit from the ssh, I get the machine back as I have left, when I reconnect. The issue is there only when the session disconnects itself. There have been instances where the session disconnected and I was able to connect back as well.
The issue is not there on all my VMs.
From the suggestions from the link I have posted above:
I am not connected to the cloud shell. i am taking ssh of the machine using the chrome extension
Have not manually mounted any disks (afaik)
I have checked the logs from gcloud compute instances get-serial-port-output --zone us-east4-c INSTANCE_NAME. I could not really make much of it. Is there anything I should look for specifically?
Any help is appreciated.
Please find the links to the logs as suggested by #W_B
Below is from 8th when the machine was restarted and files deleted
https://pastebin.com/NN5dvQMK
It happened again today. I didn't run the command immediately then. The below file is from afterwards though
https://pastebin.com/m5cgdLF6
The below one is after logout today.
[4]: https://pastebin.com/143NPatF
Please note that I have replaced the user id, system name and a lot of numeric values in general using regexp. So, there is a slight chance that the time and other values have changed. Not sure if that would be a problem.
I have added the screenshot of the current config from the UI
Using locally attached SDD seems to be the cause ... here it is explained:
https://cloud.google.com/compute/docs/disks/local-ssd#data_persistence
You need to use a "persistent disk" - else it will behave just as you describe it.
I am troubled with my in browser ssh download functionality of google cloud platform. It was working quite well until recently when I started to observe un expected behavour. Everytime I have attempted to download a file the progress bar shows as if something is downloading but after completing nothing is found in the downloads folder. I have tried to switch and log onto another os partition on my local machine. But still experience the same problem even when I changed browser program nothing seems working. Thanks in advance!
Google experienced an issue around the same time you reported the error. It has since been marked as resolved.
Are you still experiencing the same issue?
If that is the case, you can report it in the Issue Trackers platform.
I have the following problem: I have a Laptop with two different user accounts (UserA & UserB / they belong to the same person but have different priveleges).
UserA: can build the solution and run tests locally (Everything is working as expected)
UserB: Has the rights to publish to the specific network drive, however he cannot build the project. The reason is that the Windows user itself has not the correct proxy settings configured, and thats why the nuget packages cannot be downloaded.
I have no possibility to change the proxy settings for UserB right now (It depends on internal processes, which might take a while)
However I would have expected, that when I build the solution with UserA, UserB would not need to build it again and just can publish it. This, does not work, as soon as I want to publish, VS tries to rebuild the solution and this does not work because some dependencies are missing and cannot be loaded.
Is there any solution for this problem? I tried to make a research, however I was not really sure for the correct keywords to search for.
Edit: I have now the domain added to my UserB, but I still have some other problems. However I found out that this article is heavily related to my issue: Unable to launch Visual Studio 2015 as a different user.
When I start Visual Studio via Command Line with the mentioned arguments, more works, however I am now not able any longer to connect to my database using Integrated Security = true in the connection string. But at least I can build now. Deploying also works but I just get a 500 Server error when I try to connect to the resource.
Edit2: I needed to add the domain to my user when opening Visual Studio with the command mentioned in the link above. This fixed my problem with connecting to the database.
Actually I do not need an answer anymore for my initial question but I will not delete it because maybe someone has an answer for a person who will have the same question later on.
When I have the AWS Toolkit installed, it takes a long time for Visual Studio (2015 update 2) to open.
Why? Because (as I saw using Fiddler) it's posting to 169.254.169.254/latest/dynamic/instance-identity/document like 9 times, and timing out every time.
I found this article which explains what that's all about.
But how do I turn it off? I really don't want to wait for so long for Visual Studio to load every time I open it.
The underlying SDK that the toolkit uses had a bug, fixed last week, where it was attempting to always fetch instance metadata to try and determine the AWS region (even when not running on an EC2 instance). I don't recall the version of the toolkit being built on that SDK version but it's possible.
A colleague posted an updated toolkit (1.9.6.21) this evening for an unrelated change - can you install this and see if you still see an issue? If so I'll take a look when I get into the office.