I have uploaded a csv for image classification. I have checked the csv for any errors or duplicate values. But when I import the csv, i get the error that annotation is deduped and only a few images out of all the images are getting imported.
i labelled the images without whitespaces
What could be the possible cause for this?
Related
I have a problem with my CSV that is created using python. I'm following the example of google cloud about how to create JSONL from my PDF files, the big problem is that I can make my CSV with JSONL URL'S form my PDF files but this JSONL doesn't have any label. I'm using this file ->
input_helper_v2.py to create my CSV with JSONL, after that when I upload my CSV to google cloud for train my AutoML I got this error:
I tried to put some labels by myself in the CSV but that doesn't work. I don´t know if maybe its not the correct way to do this but I can find any solutions
This is an example of my CSV without labels:
Error Message: "Failed to create table: Error while reading data, error message: CSV table references column position"
I'm having issues loading data from a CSV in Google Cloud Storage into BigQuery and creating an associated table. I'm starting with Cloud Storage, adding my raw CSV file there. Then, moving to BigQuery, Create Dataset > Create Table using the CSV in the Cloud Storage.
My CSV format is;
enter image description here
The Parameters in my BigQuery table are;
enter image description here
I can't get the data to load while in this format and this setup. the original dataset goes to 10k+ plus rows, however I've reduced the scope to troubleshoot the format error.
Any response or guidance would be greatly appreciated
I configured a DLP inspection job using GCP Console to scan PDF and DOCX files. It is working as expected, finding the expected entities and saving results to a BigQuery table.
According to the docs, DLP uses Intelligent Document Parsing for PDF and DOCX. This should give me additional location details in a DocumentLocation object.
I can see column location.content_locations.document_location.file_offset in the BigQuery table, but it is empty.
I am getting location.byte_range values for TXT files and location.content_locations.image_location.bounding_boxes for images, but no location information for documents.
What can be causing this issue?
I'm trying to perform custom text classification by using AutoML on Google Cloud Platform. I am using the official google documentation to help me get started. The link to the blog is https://cloud.google.com/blog/products/ai-machine-learning/no-deep-learning-experience-needed-build-a-text-classification-model-with-google-cloud-automl-natural-language
In the above above blog they have used the 20 Newsgroup dataset. After preparing the dataset and following the instructions given here I am getting an error while uploading the dataset into GCP AutoML Text Classification GUI.
I have also tried to upload a csv file with just one data entry that also doesn't seem to work.
Every time I try to upload dataset I get the following error
ERROR CODES:
4
Last error message
CSV file is empty
More like a csv issue. If you use the csv file provided in the quickstart then it will work.
When I tring to import bigquery tables as dataset in my data Prep flow, I have the following error:
Could not create dataset: I/o error.
I tried to import many bigquery tables (which are from same BQ dataset) all of them successfully imported except this which has many columns (more than 2700 columns!).
Maybe that's because of the large number of columns but I can't see any such limitation in Docs!
when I selecet the table ----> I have this message "Preview not available" like this:
and after clicking "import":
Does anyone have any idea why this is happening or has any suggestion?
Dataprep documentation doesn't have maximum columns limitation, thus it's not probable that this is the problem.
On the other hand, I have seen generic messages like 'I/O error' or simply red icons when importing data, and they are related to the data type conversion between Dataprep and BigQuery.
I think that finding the types in BQ that are not compatibles in Dataprep and converting to one compatible should solve your issue.