How to find AWS Docker Repo - amazon-web-services

I went here and started on the first task which is to create a registry. I later closed my browser and when I go back to that page, I just get the homepage again and if I start that wizard, it acts as if I've never done it before and forces me to create a new repository.
How the heck do I get back to the repository I created initially and then how can I continue on with this wizard to the next steps with that repo? Or do I lose the repo entirely until I get through all steps in this wizard? Where the heck did my repo go? It says it exists but where? How do I get to back to that repo on the AWS control panel?

https://console.aws.amazon.com/ecs/home?region=us-east-1#/repositories
no? I couldn't comment instead of posting as answer due to insufficient reputation, sorry for that.

Related

GCP Source Repository missing from All repositories list

TL;DR
My test-repo is missing from All repositories list inside Cloud Source Repository panel, but I can still access it. Why?
Also asked here on Google Cloud Community
Details
I made test-repo
I was able to git clone using CloudSDK, make changes and even push.
When I went back to GCP Console > Cloud Source Repository, the repository did not show up and it showed me a welcome page.
I was still able to git pull.
I was able to access test-repo by going to my Cloud Function which was using the source code from test-repo.
When I made a new temporary repo caleld test2, it showed up under the All Repositories tab.
test-repo is still missing, but it shows up under Recently Viewed and I can access it there.
What is happening here???
Edit 1:
???
It is now only showing test-repo only.
Since you are able to reach your repository on every step of the way using the Cloud SDK, and even using the UI at some point, I would say this this is likely an UI issue in the GCP console.
That being said I recommend you to open an issue in Google's Issue Tracker so that Google Cloud's Engineering team can be aware of the issue and work towards fixing it.

Cloud Function build error - failed to get OS from config file for image

I'm seeing this Cloud Build error when I try to deploy a Cloud Function:
"Step #2 - "analyzer": [31;1mERROR: [0mfailed to initialize cache: failed to create image cache: accessing cache image "us.gcr.io/MY_PROJECT/gcf/us-central1/SOME_KEY/cache:latest": failed to get OS from config file for image 'us.gcr.io/MY_PROJECT/gcf/us-central1/SOME_KEY/cache:latest'"
I'm able to build and emulate the cloud function locally, but I can't deploy it due to this error. I was able to deploy just fine until now. I've looked everywhere and I can't find any discussion about this. Anyone know what's going on here?
UPDATE: I deployed a new function 3 days ago and now I can't seem to deploy an update to it. I get the same error. I'm fairly sure this is happening due to the lifecycle rule I set up to ensure I don't keep storing images of functions: Firebase storage artifacts is huge and keeps increasing. This rule is important to keep around because I don't want to pay for unnecessary storage, but it seems like it might be the source of our problem here. Can someone from Google look into this?
I got the same error, even for code that deployed successfully before.
A workaround is to delete the Docker images for the failing Firebase functions inside Container Registry and re-deploying the functions. (The images will be re-created upon deploying.)
The error still occurs sporadically, so I suspect this may be a bug introduced in Firebase's deployment process. Thankfully for now, the workaround above resolves the issue every time the error comes up.
I also encountered the same problem, and solved it by deleting the images in the Container Registry of Firebase Project.
I made a Script at that time, and I'll put it here. The usage is as follows. Please use it if you like.
Install the Google Cloud SDK.
Download the Script
Edit CONTAINER_REGISTRY to your registry name. For example: CONTAINER_REGISTRY=asia.gcr.io/project-name/gcf/asia-northeast1
Grant execute permission. - $ chmod +x script.sh
Execute it. - $ sh script.sh
Deploy your functions.
I'm having the same problem for the last few days and in contact with the support. I had the same log and in my case it wasn't connected to the artifacts because the artifacts rebuild themselves automatically on deploy (read below about a subtle case related to the artifacts and how to fix it), but deleting the functions and redeploying solved it for me.
Artifacts auto cleanup
Note that if the artifacts bucket is empty, then the problem is somewhere else.
But if it's not empty, what you can do to resolve any possible problems related to the artifacts auto cleanup, is to delete the whole "container" folder manually in the artifacts which should solve it. Then just redeploy again.
Make sure not to delete the artifacts bucket itself!
Dough from firebase confirmed in the question you referring to that removing the artifacts content is safe.
So, here is how to delete it:
go to the google cloud console, select your project -> storage -> browser https://console.cloud.google.com/storage/browser
Select the "artifacts" bucket
Choose "containers" and delete it
If the problem was here, it should work fine after that.
This happens because the deletion rule you refer to in your question checks the "last updated" timestamp of each file while on redeploy only some files are updated. So the next day the rule will delete some of the files while leaving the others which will lead to the inconsistent state of the bucket in this case. So you just remove everything manually.

Django/Heroku After pushing a new version certain templates do not update

I have made some changes to a Django template, saved them, committed them to git, pushed to GitHub and then deployed this new version to Heroku.
The build succeeds and I can view the template but the change is not there.
I checked the GitHub repo, the change is there. When I check the activity feed on Heroku I can see the change when I click the Compare diff link.
When I make changes to other areas of the project, views etc there's no problem, I have updated other templates without issue but can not get the new CSS into the deployed version.
Inspecting the code of the deployed version shows the old CSS not the new so it isn't just an issue with my stylesheet, inheritance or so on.
I've run out of ideas, any suggestions of how I can resolve this?
I have found an answer, it seems a little like using a sledge hammer to crack a nut but ...
Anyway I deleted the templates directory, went through a cycle of commit, push deploy. Reinstated the templates directory redid the commit, push, deploy cycle and for whatever reason the CSS is now being picked up and displayed as expected.
I still have no idea what the original problem was, why it happened or how probable it is I will have the same sort of issues again so any suggestions, explanations and clarifications would still be much appreciated.

How do you maintain the manual setup of Drupal installation?

I have set up an elastic beanstalk deployment of Drupal to host a Drupal built website.
When I start up my ec2 instance, I go through the installation steps of setting up Drupal.
However, when the instance is Restarted, or Stopped, restarting the instance goes back to this page!
How can I configure the instance so that these installation steps do not need to be repeated even when the instance goes down. This is quite worrying as I am looking to host my website this way.
Any help on this issue would be greatly appreciated!
This step means that drupal can't find database or it can find it but installation is not done yet (required tables are not created). So if you pass installation and you see again this screen it can mean 2 things:
Your database configuration is lost, so drupal can't find DB.
DB configuration file is ok, but DB it self is lost so configuration must be done again.
So first check what is the problem in your case and then solve it. Probably since you are installing in instance, DB configuration is then recorded and then lost when your box expires. If so...find a way to make permanent change do config file (should be /sites/default/settings.php).

Malware/Adware on site deployed on AWS Elasticbeanstalk

I guess the title says it all.
I noticed some ads were popping up on a client's site we are currently developing. It only shows on this particular site. Not any other site. It is very annoying to put it mildly.
I thought removing it would be as easy as setting up a new environment on ElasticBeanstalk for it. I was wrong!
I have started a fresh instance for the application, scanned the project folder for malware before deploying, emptied the content of s3 bucket for static files. All these made no difference. The adware/malware is still there.
It has been driving me nuts for the past few days. Does anyone know how to resolve this kind of problem?
Mark B pointed me in the right direction.
I used inspect element to check the network processes of pages showing the Malware/Adware. It was after this i noticed a few asynchronous posts going to http://api.adsrun.net/post. Of course, i'm not making any post calls to this link. So i decided to inspect my JavaScript files as seen in View Page Source. Fortunately, it was in the last few lines of the second file i inspected. Immediately i deleted this file, normalcy returned to my web application.
It has been a very frustrating several hours. Thanks once again, Mark B for your suggestion.