Importing and parsing a large CSV file with go and app engine's datastore -


Locally I successfully (in action):

  • Open the CSV

  • Scan through each line (Scanner.
  • Map the desired structure to the parseed CSV line
  • Save the Stret to the datastore

I think that I Will allow - but there is a limit of 32MB. I also see that there is a bulk upload tool - bulk_uploader.py - but not all data-mastered Sing what I want and I want to limit the type of bulk entry (and the cost).

How would a very large (500 MB +) CSV file read and parse effectively without the benefit of reading from local storage?

You have to look at the following options and see if it works for you:

< Ol>
  • Looking at the larger file size, you should consider using it for a file. You can use command line utilities which GCS makes available to upload your file to your bucket. Once uploaded, you can view directly using the JSON API to work with the file and import it into your datastore layer. Take a look at the following:

  • If it is like importing a large file once, then another option can spin a Google Compute VM, in which to write an app Read from GCS and pass on data through small parts for service running in AP engine, which can then accept and continue the data.


  • Comments

    Popular posts from this blog

    java - org.apache.http.ProtocolException: Target host is not specified -

    java - Gradle dependencies: compile project by relative path -

    ruby on rails - Object doesn't support #inspect when used with .include -