I'm using Wowza for recording a live stream. I want to get a thumbnail of the live stream. I followed this tutorial
http://blog.svnlabs.com/wowza-live-stream-thumbnail-images-with-http-provider/
Then when I record stream, it saved about 4 video file:
streamname.mp4,
streamname_160p.mp4,
streamname_320p.mp4,
streamname_source.mp4
Is there any way to save only 1 video file "streamname.mp4" but could still get the thumbnail by URL as the above tutorial mentions?
Thanks for all your help!
There is no way.
We have to remain at least one of extending file
e.g
streamname.mp4,
streamname_160p.mp4,
Related
I am a new Requestor, and trying to create a new HIT based on the Audio Naturalness HTML template in https://requester.mturk.com/create/projects/new.
I would like to use wav files, and in the html file provided, I have modified the audio_url line to look like:
<!-- Your audio file URLs will be substituted for the "audio_url" variable when you publish a batch with a CSV input file containing multiple
audio file URLs -->
<source src="${audio_url}" type="audio/wav" />
When I'm going to publish the batch, it asks for a csv, which looks like:
audio_url
https://github.com/user/voiceTTS/blob/master/Sample1a.wav
https://github.com/user/voiceTTS/blob/master/Sample1b.wav
However, when I go to preview the HIT, the audio files are blank, i.e. nothing plays.
Where am I going wrong?
This was actually an issue with hosting audio files on GitHub. I hosted them on my personal website, and it works as expected!
I want to download a particular no. of videos from a Youtube playlist where they are arranged one after the other. How can I do so using youtube-dl and FFmpeg? Also, I want to number them according to their numbers in the playlist.
Use --playlist-end and a template with playlist_index, like this:
youtube-dl --playlist-end 2 -o "%(playlist_index)s-%(title)s-%(id)s.%(ext)s" PLiZxWe0ejyv9R8q1jN70HZ4mwhTJxSIXf
Replace 2 with the number of videos you want to download, and PLiZxWe0ejyv9R8q1jN70HZ4mwhTJxSIXf with the playlist ID or playlist URL you actually want to download.
I'd like to ask if there is any way to get thumbnail image for media returned from Instagram Graph API? I can get an image URL by using following endpoint: /{InstagramUserId}/media?fields=media_url
However it only returns one size. Old Instagram API returned various sizes like low_resolution, thumbnail, standard_size. Is it possible to get similar result by using Instagram Grahp API?
After searching for some time to a solution i finally found one.
So, include in the fields query string the permalink field and it should give as following (Sorry, for using Kim Kardashian as an example):
https://www.instagram.com/p/CAYDz52gSLh
then append this at the end media?size=t or simply media like this:
https://www.instagram.com/p/CAYDz52gSLh/media?size=t
media?size=t will give a 150x150 px image
media?size=m will give a 320x320 px image
media?size=l will give a 1080x1080 px image
I can not find a way to get the thumbnails in one step.
But in the next step after receiving the data, the thumbnail of each image and video can be found in this way using oEmbed:
https://graph.facebook.com/v9.0/instagram_oembed?url={permalink}&maxwidth=320&fields=thumbnail_url,author_name,provider_name,provider_url&access_token={accessToken}
permalink: The url received from the previous stage (business_discovery)
Example:
https://graph.facebook.com/v9.0/instagram_oembed?url=https://www.instagram.com/p/CBOuPY1AcMQ/&maxwidth=320&fields=thumbnail_url,author_name,provider_name,provider_url&access_token={accessToken}
More Details
I have my own image encoder and decoder. What I want is to read DICOM images, extract uncompressed image pixels, compress them using my encoder and then write those compressed pixels back in the DICOM file in place of uncompressed pixels. Decoder would do the opposite. Is this possible to do in Imebra? I can read tags and pixels, but after I compress them, I'm not sure how to put them back (they are currently in char* buffer), or if this is even possible. I'm using Imebra in C++.
Alternatively, it would be okay if I could create completely new DICOM file, but in that case I would need to easily transfer all the DICOM tags from the old file.
If this is not possible in Imebra, is there some other C++ library that allows this?
Edit:
Thanks for the answer, Paolo. However, original DICOM image still remains unchanged (using second option). Can you say am I doing something obviously wrong here?
std::unique_ptr<imebra::DataSet> loadedDataSet(imebra::CodecFactory::load(imgNameM));
imebra::WritingDataHandlerNumeric* dataHandler = loadedDataSet->getWritingDataHandlerRaw(imebra::TagId(imebra::tagId_t::PixelData_7FE0_0010), 0);
dataHandler->assign(buffer, size);
delete dataHandler;
loadedDataSet is not empty, I checked with bufferExists.
Edit 2:
Yes, I didn't save it. Now I added that line and managed to modify PixelData element which was solves my original problem. Thanks. However, some other parts of the file are now automatically also changed. More than 100 empty bytes are added at the beginning of the file, although this doesn't bother me that much. What bothers me is that (0008,0005) Specific Character Set tag is now added and its value isn't set (it's empty) which then causes CharsetConversionNoTableError when trying to read tags of that modified file. If I remove that tag manually and fix the group length, I can read tags normally. Is there a possibility to avoid this behavior?
Imebra already provides the encoder/decoder for lossless jpeg, baseline and extended jpeg, RLE.
There are several ways of adding your own codec to Imebra:
derive a class from imebra::implementation::codecs::imageCodec
or encode an image into a char buffer, then add it as raw content using imebra::DataSet::getWritingDataHandlerRaw which allows you to write the tag raw content. getWritingDataHandlerRaw returns a WritingDataHandlerNumeric. Use WritingDataHandlerNumeric::assign to move the bytes buffer into the data handler, then delete the data handler to cause it to commit its content into the dataset.
The example changeTransferSyntax that comes with the library shows how to create a new dataset with all the same tags from the source dataset but with a different transfer syntax (including a different image compression)
I try to compose new and modify existing multipage TIFFs using Magick++.
Does someone know how I can do this?
I can read a specific page using this code:
Image * img = new Image("path/to/image.tif[0]"); //read page 0
But how can I save changes back to the TIFF? and how can I add new pages?
Google could only tell me things about splitting TIFFs in singe page ones.
Thanks for your help!
I never tried it but what about the writeImages function. Docs are here.
From the docs
Write images in container to file specified by string imageSpec_