Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 13 hours ago.
Improve this question
Despite many attempts, I can't seem to make any adobe EdgeAnimate composition work on Amazon S3. I appreciate a website hosted on S3 must be static, but looking through the code exported by EdgeAnimate, no sever side aid is needed. It's just HTML, JS, CSS files, all of which I use in websites that work without problems on S3.
All files are of course publicly accessible and the bucket is configured as a website.
Perhaps a CORRS config or an included XML file may fix it, but I've no experience with either (same goes for PHP, SQL etc..hence the need to run off S3).
I've posed the same question on AWS and Adobe forums, to no avail, so here goes a final attempt. Any suggestions will be gratefully received.
Thanks in advance.
Forum Links -
Amazon - https://forums.aws.amazon.com/thread.jspa?threadID=143338
Adobe - http://forums.adobe.com/thread/1373062
Related
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed last year.
Improve this question
I have a personal website hosted on aws using s3, cloudfront and route53, but as time goes by I would like to have some dynamic functionalities such as allowing visitors to leave a private message. May I know what is the best option for me? Should I use an EC2 instance? Or Lambda + API gateway... etc?
Any suggestion is appreciated, thanks!
I would recommend that you start at Choose Your Web Application Infrastructure because there is no single answer.
There are different factors to take into account when deciding which service to use to deploy your first web application on the AWS Cloud. You should consider if you want a managed infrastructure with low stress, if you need containers (and why) or if you want full control and customization.
As I think you currently don't have server-side rendering type of setup that is great news for you to host a personal website, For any dynamic backend related stuff you can use AWS Lambda functions for contact forms submissions and other things and the price would also no very low to host these rather than switching to ec2
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 2 years ago.
Improve this question
I was doing one of those self-paced tutorials that were offered when I first logged it to guide me through loading data into a storage bucket and copying it into BigQuery from there and so on. I refreshed the page and it went away and I have not been able to find a way to get it back. Any idea how to get back to a tutorial I was working through?
You can find the quickstart guides for BigQuery in the official GCP documentation.
Here, you can also find all how-to guides, related to the BigQuery. For loading data from Cloud Storage into BigQuery in particular, check the section called "Loading data into BigQuery" - link.
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 4 years ago.
Improve this question
I was discussing with friends about the best way to store files in amazon.
I believed that an s3 bucket was the best way to save static files as images from a website.
But friends said that it is not the best way to work with s3 because of the high cost of having this image requested many times.
I need to know the best place to save images that will be rendered inside my site (which is inside an EC2 instance).
Could someone clarify this doubt? Saving images inside the S3 in sites that have many requests is expensive?
For storing static files like images AWS S3 is one of the best option.
S3 is one of the cheapest cloud storage, you won't be charged for the number of times it's read, only amount of outbound traffic will bbe charged. For get requuest/put request there is also a charge, but you shouldn't need it as per my understanding, you can clarify your use case more precisely. You can also calculate the price here.
Find all the storage services AWS offers here: https://aws.amazon.com/products/storage/
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I read in reddit wiki that reddit moved to indextrunk but when I reviewed run.py file I found that there is keys like Cloud_Search_Api_key ... So I guessed it is using Amazon cloud search . If this true what are the values that should be changed in run.py to make cloudsearch works? and what is subreddit_cloud_api_key?
Thanks
I'm pretty sure Reddit uses Cloudsearch. Their Github FAQ claims they use Indextank, but IndexTank has been shut down since April 2012. If you search on Reddit and highlight the "δ" symbol, it will show text like "δ converted query to cloudsearch syntax: (and (field text 'search') (field text 'terms'))".
I'm not too familiar with Python or AWS, but it looks like CLOUDSEARCH_SEARCH_API and the other similar variables are URLs that Amazon calls Endpoints.
The variable names in reddit/r2/run.ini contain SEARCH and DOC, mirroring Amazon's documentation. Also, cloudsearch.py makes an HTTP connection to that variable:
search_api = g.CLOUDSEARCH_SEARCH_API
//...
connection = httplib.HTTPConnection(search_api, 80)
So you would probably set CLOUDSEARCH_SEARCH_API with the URL to your Cloudsearch endpoint.
EDIT: Kemitche has answered this on Reddit. Unlike me, he knows what he's talking about, so take a look.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 3 years ago.
Improve this question
I am looking to build a webapp to improve user experience in booking railway tickets in India. The API is impossible to get due to hefty charge to procure it. I have seen many apps that provide details of the trains etc through their apps.
My Question is how are they scraping data from the website.In general how can I legally get data shown to user (I don't want payment and stuff that are impossible without API) on any website. How do people scrape such data? Any tools/methods?
Bear with me if question is naive. I'm pretty new to this stuff.
They can get the train schedule information using any one of several programming languages though it is most likely done with ordinary PHP and any good webserver host. For example all indian train schedules can be found on the indianrail.gov website.
Sending a specially built URL to ..
http://www.indianrail.gov.in/cgi_bin/inet_trnnum_cgi.cgi?lccp_trnname=1123
using the POST method of sending form data should give you all the details for train number 1123 After that it becomes just a simple task of tidying up the results for storage in a database.
Update: well armoured site its checking both the user agent and referer of inbound requests.
Ammendum: the indianrail.gov site is changing to http://www.trainenquiry.com/ -> will have to take another look