I currently have the problem that I can not upload images to person models using the Video indexer website anymore. Person models as well as the individual persons can be easily created and added.
But when I try to add images, no images are loaded.
The message "0 images uploaded" appears and does not change with time.I have already tried it with images that have already been successfully uploaded to other person models, but unfortunately it was also unsuccessful.
Does anyone perhaps have a solution for this?
Thanks for contacting us! I'm from Video Analyzer for Media team (former known as Video Indexer).
This is a bug that we are aware of and already been resolved.
It is currently in deployment to all regions, it will be available in the next 24hr.
Related
I am currently developping an Flask application that dynamically generated images. I save the image to static/img folder.
But the image is never changed after first time created.
Any body know what the issue behind.
many thanks.
It could be a caching issue (especially since you're saving to a static folder). Try appending a dummy query parameter e.g <your_url>/?123. If you see the new file, then it's a caching issue. One quick and dirty fix would be to generate unique values and append to the url or you can look up cache bursting techniques for GAE
I've already been granted by the ImageNet website http://www.image-net.org/download-images to download the image data. And the page shows:
You have been granted access to the the whole ImageNet database through our site. By doing so you agree to the terms of access.
Download as one tar file
The full ImageNet data is currently unavailable. Data for ILSVRC is available.
ImageNet Fall 2011 release MD5: ...
ImageNet10K from Deng et al. ECCV2010
But both of the links shows "OOPS The url is not valid." when opened. (It's absolutely not due to some problem of my web or browser. I can tell this by the consistency of ImageNet web page style. I guess these links are too old, and moved to other urls, yet their website didn't update at once)
I have two questions here.
(1) Where and how can I really download the ImageNet data (as well as their labels, for classification task)?
(2) I want the data for the purpose to validate my method in my paper. Even if the dataset is downloaded, I'm afraid that it's unnecessarily big. Do I have to validate on ImageNet (Since its adopted in many papers anyway...) ? The Tiny ImageNet data's page seems not broken on their website. But its a dataset much smaller.
ImageNet Download:
Go to https://www.kaggle.com/c/imagenet-object-localization-challenge and click on the data tab. You can use the Kaggle API to download on a remote computer, or that page to download all the files you want directly.
There, they provide both the labels and the image data.
I don't know what is up with the ImageNet website, however, the url list links were also broken for me today. One way you can still get the data is by going to an alternate mirror, such as Kaggle ImageNet download, the link I provided above. From what I have hears, the Kaggle ImageNet is equivalent to the ImageNet from their website.
I'm unsure about how to answer your second question, as I don't know enough about your project. However, ImageNet will probably work to validate your model.
It can be downloaded in python using the datasets library:
>>> from datasets import load_dataset
>>> ds = load_dataset("imagenet-1k")
>>> train_ds = ds["train"]
>>> train_ds[0]["image"] # a PIL Image
You may need to install it as well as Pillow and login to Hugging Face after accepting the terms of access
pip install datasets Pillow
huggingface-cli login
You can find more info and links to download the files on the ImageNet page on Hugging Face: https://huggingface.co/datasets/imagenet-1k
I've looked for this across the web a few times, and I feel like this hasn't been asked exactly, or I may just be getting bogged down with the wrong syntax. Hoping to get an easy answer here (yes, you can't get this, is an acceptable answer).
The variations from the base CentOS image are listed here: Link to GCP
However, they don't actually provide a download for this image. I'm trying to get a local VM running in VMWare with this image.
I feel as though they'd provide this to their clients to make it easier to prepare for use of their product, but I'm not finding it anywhere.
If anyone could toss me a link to a pre-configured CentOS ISO with the minor changes, I'd definitely take that as an alternative. I'm just not confident in my skills with Linux enough to configure the firewall properly :)
GCP doesn't support Google-provied images for exporting. However, they support exporting images for custom images.
I don't have any experience about image exporting, but I think this works.
Create custom images
You can create custom images based on your GCE VM instance.
Go navigation -> Compute engine -> images page.
You can create custom image via disk or snapshot in this page.
Select one and create a custom image.
Export your image
After creating custom image successfully, Go custom image page and click "export" on upper side.
Select export format and GCS destination. then click export.
Now you have an image in the Google Cloud storage.
Download image file and import to your local VM machine.
I want to store the images along with the thumbnails of it. I am storing the images in file system using django. At first, the user will be able to see the thumbnails and after clicking it original images can be seen. I am using postgres database. Also, I have already installed Pillow library. Thumbnail size will be approx 200*200.
Now my questions are:
How should I store the thumbnails ? (in database or in file system)
How to convert the images to it's thumbnails ? (python library or something else)
If anything better is possible for the mentioned feature please let me know.
P.S.: High performance and lesser page load time is required.
There are third party apps that do the heavy lifting like sorl-thumbnail or easy-thumbnail
For the first question, storing the image in the system or cdn and the path in the db is the best approach, and that's what django does by default.
Am working on an aldryn-newsblog. it works well on my local host.
once i push the blog to heroku, thats when i get a problem.
Every time i add a new article, the article publishes very well with no problem. but after some time,say 20 mins the featured image just disappears. This forces me to edit the article again so as to add the image afresh.
image showing my file missing once i try to edit my article
What might be the problem? kindly assist.
Heroku uses ephemeral file system (files may be written to the app’s “tmp/” directory, but can be discarded by the system at any time)
Use a cloud storage service instead, see https://devcenter.heroku.com/articles/s3
Heroku docs: https://devcenter.heroku.com/articles/dynos#ephemeral-filesystem