Load huge image dataset and some data augmentation using Pytorch












0














I have a question. I have about 2 million images (place365-standard dataset) and I want to do some data augmentation like transforming, cropping etc. Also, I have to make my own target image (y) based on some color model algorithms (CMYK) for example.



So Actually, my preprocessing step includes augmentation and making terget image (y). And then I should feed these images to a deep network. When should I do this based on Dataset class? Should I do my processing step in __getitem__()? If yes, would it be parallel and fast?



Here is my template of Dataset(data.Dataset) class:



import torch
from torch.utils import data

class Dataset(data.Dataset):
"""
Return Dataset class representing our data set
"""
def __int__(self, list_IDs, labels):
"""
Initialize data set as a list of IDs corresponding to each item of data set and labels of each data

Args:
list_IDs: a list of IDs for each data point in data set
labels: label of an item in data set with respect to the ID
"""

self.labels = labels
self.list_IDs = list_IDs

def __len__(self):
"""
Return the length of data set using list of IDs

:return: number of samples in data set
"""
return len(self.list_IDs)

def __getitem__(self, item):
"""
Generate one item of data set. Here we apply our preprocessing things like halftone styles and subtractive color process using CMYK color model etc. (See the paper for operations)

:param item: index of item in IDs list

:return: a sample of data
"""
ID = self.list_IDs[item]

# Code to load data
X = None #

# code to apply your custom function to make y image (time consuming task - some algorithms)
y = None #

return X, y


Thanks for any advice



Best regards










share|improve this question





























    0














    I have a question. I have about 2 million images (place365-standard dataset) and I want to do some data augmentation like transforming, cropping etc. Also, I have to make my own target image (y) based on some color model algorithms (CMYK) for example.



    So Actually, my preprocessing step includes augmentation and making terget image (y). And then I should feed these images to a deep network. When should I do this based on Dataset class? Should I do my processing step in __getitem__()? If yes, would it be parallel and fast?



    Here is my template of Dataset(data.Dataset) class:



    import torch
    from torch.utils import data

    class Dataset(data.Dataset):
    """
    Return Dataset class representing our data set
    """
    def __int__(self, list_IDs, labels):
    """
    Initialize data set as a list of IDs corresponding to each item of data set and labels of each data

    Args:
    list_IDs: a list of IDs for each data point in data set
    labels: label of an item in data set with respect to the ID
    """

    self.labels = labels
    self.list_IDs = list_IDs

    def __len__(self):
    """
    Return the length of data set using list of IDs

    :return: number of samples in data set
    """
    return len(self.list_IDs)

    def __getitem__(self, item):
    """
    Generate one item of data set. Here we apply our preprocessing things like halftone styles and subtractive color process using CMYK color model etc. (See the paper for operations)

    :param item: index of item in IDs list

    :return: a sample of data
    """
    ID = self.list_IDs[item]

    # Code to load data
    X = None #

    # code to apply your custom function to make y image (time consuming task - some algorithms)
    y = None #

    return X, y


    Thanks for any advice



    Best regards










    share|improve this question



























      0












      0








      0


      3





      I have a question. I have about 2 million images (place365-standard dataset) and I want to do some data augmentation like transforming, cropping etc. Also, I have to make my own target image (y) based on some color model algorithms (CMYK) for example.



      So Actually, my preprocessing step includes augmentation and making terget image (y). And then I should feed these images to a deep network. When should I do this based on Dataset class? Should I do my processing step in __getitem__()? If yes, would it be parallel and fast?



      Here is my template of Dataset(data.Dataset) class:



      import torch
      from torch.utils import data

      class Dataset(data.Dataset):
      """
      Return Dataset class representing our data set
      """
      def __int__(self, list_IDs, labels):
      """
      Initialize data set as a list of IDs corresponding to each item of data set and labels of each data

      Args:
      list_IDs: a list of IDs for each data point in data set
      labels: label of an item in data set with respect to the ID
      """

      self.labels = labels
      self.list_IDs = list_IDs

      def __len__(self):
      """
      Return the length of data set using list of IDs

      :return: number of samples in data set
      """
      return len(self.list_IDs)

      def __getitem__(self, item):
      """
      Generate one item of data set. Here we apply our preprocessing things like halftone styles and subtractive color process using CMYK color model etc. (See the paper for operations)

      :param item: index of item in IDs list

      :return: a sample of data
      """
      ID = self.list_IDs[item]

      # Code to load data
      X = None #

      # code to apply your custom function to make y image (time consuming task - some algorithms)
      y = None #

      return X, y


      Thanks for any advice



      Best regards










      share|improve this question















      I have a question. I have about 2 million images (place365-standard dataset) and I want to do some data augmentation like transforming, cropping etc. Also, I have to make my own target image (y) based on some color model algorithms (CMYK) for example.



      So Actually, my preprocessing step includes augmentation and making terget image (y). And then I should feed these images to a deep network. When should I do this based on Dataset class? Should I do my processing step in __getitem__()? If yes, would it be parallel and fast?



      Here is my template of Dataset(data.Dataset) class:



      import torch
      from torch.utils import data

      class Dataset(data.Dataset):
      """
      Return Dataset class representing our data set
      """
      def __int__(self, list_IDs, labels):
      """
      Initialize data set as a list of IDs corresponding to each item of data set and labels of each data

      Args:
      list_IDs: a list of IDs for each data point in data set
      labels: label of an item in data set with respect to the ID
      """

      self.labels = labels
      self.list_IDs = list_IDs

      def __len__(self):
      """
      Return the length of data set using list of IDs

      :return: number of samples in data set
      """
      return len(self.list_IDs)

      def __getitem__(self, item):
      """
      Generate one item of data set. Here we apply our preprocessing things like halftone styles and subtractive color process using CMYK color model etc. (See the paper for operations)

      :param item: index of item in IDs list

      :return: a sample of data
      """
      ID = self.list_IDs[item]

      # Code to load data
      X = None #

      # code to apply your custom function to make y image (time consuming task - some algorithms)
      y = None #

      return X, y


      Thanks for any advice



      Best regards







      python image-processing deep-learning bigdata pytorch






      share|improve this question















      share|improve this question













      share|improve this question




      share|improve this question








      edited Nov 24 '18 at 6:40







      M. Doosti Lakhani

















      asked Nov 23 '18 at 22:52









      M. Doosti LakhaniM. Doosti Lakhani

      324317




      324317
























          1 Answer
          1






          active

          oldest

          votes


















          1














          If you look at e.g., torchvision.dataset.ImageFolder you'll see that it works quite similar to your design: the class has transform member that lists all sorts of augmentations (resizing, cropping, flipping etc.) and these are carried out on the images in the __getitem__ method.

          Regarding parallelism, the Dataset itself is not parallel, but the DataLoader can be (see num_workers argument), so if you use your dataset inside a parallel dataloader you get the parallelism for free, Cool!






          share|improve this answer





















          • Thank you. But i have a problem. I have a 101GB .tar file contains about 1.8 million images(place365-standard), now I have to do some preprocessing like you mentioned (transforms). But I have to do something else here. I should apply some algorithms to generate new images as ground-truth (y). But I do not know where should I put this codes. Should I put my codes in __getitem__() function? Actually, I should consider RAM size, and I should do everything parallel (on GPU and CPU). I do not anything and I am new to this type of implementation.
            – M. Doosti Lakhani
            Nov 24 '18 at 18:32










          • @M.DoostiLakhani how does y relates to x?
            – Shai
            Nov 24 '18 at 18:33










          • I have to do some color model algorithms like CMYK to generate halftone image for printing or scanning applications.
            – M. Doosti Lakhani
            Nov 24 '18 at 19:14











          Your Answer






          StackExchange.ifUsing("editor", function () {
          StackExchange.using("externalEditor", function () {
          StackExchange.using("snippets", function () {
          StackExchange.snippets.init();
          });
          });
          }, "code-snippets");

          StackExchange.ready(function() {
          var channelOptions = {
          tags: "".split(" "),
          id: "1"
          };
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function() {
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled) {
          StackExchange.using("snippets", function() {
          createEditor();
          });
          }
          else {
          createEditor();
          }
          });

          function createEditor() {
          StackExchange.prepareEditor({
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: true,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: 10,
          bindNavPrevention: true,
          postfix: "",
          imageUploader: {
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          },
          onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          });


          }
          });














          draft saved

          draft discarded


















          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53453663%2fload-huge-image-dataset-and-some-data-augmentation-using-pytorch%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown

























          1 Answer
          1






          active

          oldest

          votes








          1 Answer
          1






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes









          1














          If you look at e.g., torchvision.dataset.ImageFolder you'll see that it works quite similar to your design: the class has transform member that lists all sorts of augmentations (resizing, cropping, flipping etc.) and these are carried out on the images in the __getitem__ method.

          Regarding parallelism, the Dataset itself is not parallel, but the DataLoader can be (see num_workers argument), so if you use your dataset inside a parallel dataloader you get the parallelism for free, Cool!






          share|improve this answer





















          • Thank you. But i have a problem. I have a 101GB .tar file contains about 1.8 million images(place365-standard), now I have to do some preprocessing like you mentioned (transforms). But I have to do something else here. I should apply some algorithms to generate new images as ground-truth (y). But I do not know where should I put this codes. Should I put my codes in __getitem__() function? Actually, I should consider RAM size, and I should do everything parallel (on GPU and CPU). I do not anything and I am new to this type of implementation.
            – M. Doosti Lakhani
            Nov 24 '18 at 18:32










          • @M.DoostiLakhani how does y relates to x?
            – Shai
            Nov 24 '18 at 18:33










          • I have to do some color model algorithms like CMYK to generate halftone image for printing or scanning applications.
            – M. Doosti Lakhani
            Nov 24 '18 at 19:14
















          1














          If you look at e.g., torchvision.dataset.ImageFolder you'll see that it works quite similar to your design: the class has transform member that lists all sorts of augmentations (resizing, cropping, flipping etc.) and these are carried out on the images in the __getitem__ method.

          Regarding parallelism, the Dataset itself is not parallel, but the DataLoader can be (see num_workers argument), so if you use your dataset inside a parallel dataloader you get the parallelism for free, Cool!






          share|improve this answer





















          • Thank you. But i have a problem. I have a 101GB .tar file contains about 1.8 million images(place365-standard), now I have to do some preprocessing like you mentioned (transforms). But I have to do something else here. I should apply some algorithms to generate new images as ground-truth (y). But I do not know where should I put this codes. Should I put my codes in __getitem__() function? Actually, I should consider RAM size, and I should do everything parallel (on GPU and CPU). I do not anything and I am new to this type of implementation.
            – M. Doosti Lakhani
            Nov 24 '18 at 18:32










          • @M.DoostiLakhani how does y relates to x?
            – Shai
            Nov 24 '18 at 18:33










          • I have to do some color model algorithms like CMYK to generate halftone image for printing or scanning applications.
            – M. Doosti Lakhani
            Nov 24 '18 at 19:14














          1












          1








          1






          If you look at e.g., torchvision.dataset.ImageFolder you'll see that it works quite similar to your design: the class has transform member that lists all sorts of augmentations (resizing, cropping, flipping etc.) and these are carried out on the images in the __getitem__ method.

          Regarding parallelism, the Dataset itself is not parallel, but the DataLoader can be (see num_workers argument), so if you use your dataset inside a parallel dataloader you get the parallelism for free, Cool!






          share|improve this answer












          If you look at e.g., torchvision.dataset.ImageFolder you'll see that it works quite similar to your design: the class has transform member that lists all sorts of augmentations (resizing, cropping, flipping etc.) and these are carried out on the images in the __getitem__ method.

          Regarding parallelism, the Dataset itself is not parallel, but the DataLoader can be (see num_workers argument), so if you use your dataset inside a parallel dataloader you get the parallelism for free, Cool!







          share|improve this answer












          share|improve this answer



          share|improve this answer










          answered Nov 24 '18 at 18:03









          ShaiShai

          69.3k22135242




          69.3k22135242












          • Thank you. But i have a problem. I have a 101GB .tar file contains about 1.8 million images(place365-standard), now I have to do some preprocessing like you mentioned (transforms). But I have to do something else here. I should apply some algorithms to generate new images as ground-truth (y). But I do not know where should I put this codes. Should I put my codes in __getitem__() function? Actually, I should consider RAM size, and I should do everything parallel (on GPU and CPU). I do not anything and I am new to this type of implementation.
            – M. Doosti Lakhani
            Nov 24 '18 at 18:32










          • @M.DoostiLakhani how does y relates to x?
            – Shai
            Nov 24 '18 at 18:33










          • I have to do some color model algorithms like CMYK to generate halftone image for printing or scanning applications.
            – M. Doosti Lakhani
            Nov 24 '18 at 19:14


















          • Thank you. But i have a problem. I have a 101GB .tar file contains about 1.8 million images(place365-standard), now I have to do some preprocessing like you mentioned (transforms). But I have to do something else here. I should apply some algorithms to generate new images as ground-truth (y). But I do not know where should I put this codes. Should I put my codes in __getitem__() function? Actually, I should consider RAM size, and I should do everything parallel (on GPU and CPU). I do not anything and I am new to this type of implementation.
            – M. Doosti Lakhani
            Nov 24 '18 at 18:32










          • @M.DoostiLakhani how does y relates to x?
            – Shai
            Nov 24 '18 at 18:33










          • I have to do some color model algorithms like CMYK to generate halftone image for printing or scanning applications.
            – M. Doosti Lakhani
            Nov 24 '18 at 19:14
















          Thank you. But i have a problem. I have a 101GB .tar file contains about 1.8 million images(place365-standard), now I have to do some preprocessing like you mentioned (transforms). But I have to do something else here. I should apply some algorithms to generate new images as ground-truth (y). But I do not know where should I put this codes. Should I put my codes in __getitem__() function? Actually, I should consider RAM size, and I should do everything parallel (on GPU and CPU). I do not anything and I am new to this type of implementation.
          – M. Doosti Lakhani
          Nov 24 '18 at 18:32




          Thank you. But i have a problem. I have a 101GB .tar file contains about 1.8 million images(place365-standard), now I have to do some preprocessing like you mentioned (transforms). But I have to do something else here. I should apply some algorithms to generate new images as ground-truth (y). But I do not know where should I put this codes. Should I put my codes in __getitem__() function? Actually, I should consider RAM size, and I should do everything parallel (on GPU and CPU). I do not anything and I am new to this type of implementation.
          – M. Doosti Lakhani
          Nov 24 '18 at 18:32












          @M.DoostiLakhani how does y relates to x?
          – Shai
          Nov 24 '18 at 18:33




          @M.DoostiLakhani how does y relates to x?
          – Shai
          Nov 24 '18 at 18:33












          I have to do some color model algorithms like CMYK to generate halftone image for printing or scanning applications.
          – M. Doosti Lakhani
          Nov 24 '18 at 19:14




          I have to do some color model algorithms like CMYK to generate halftone image for printing or scanning applications.
          – M. Doosti Lakhani
          Nov 24 '18 at 19:14


















          draft saved

          draft discarded




















































          Thanks for contributing an answer to Stack Overflow!


          • Please be sure to answer the question. Provide details and share your research!

          But avoid



          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.


          To learn more, see our tips on writing great answers.




          draft saved


          draft discarded














          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53453663%2fload-huge-image-dataset-and-some-data-augmentation-using-pytorch%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          A CLEAN and SIMPLE way to add appendices to Table of Contents and bookmarks

          Calculate evaluation metrics using cross_val_predict sklearn

          Insert data from modal to MySQL (multiple modal on website)