Code Monkey home page Code Monkey logo

Comments (5)

chen3082 avatar chen3082 commented on August 16, 2024

Hey men, I encounter the same issue. Are you able to resolve it?
I keep getting this,
OP_REQUIRES failed at cwise_ops_common.cc:70 : Resource exhausted: OOM when allocating tensor with shape[768,768].
I already reduce the batch size to 3 but didn't work.

from bert4doc-classification.

xuyige avatar xuyige commented on August 16, 2024

Hi,

first, thank u for having sharing ur cod with us

I am trying to further pretraining a bert model on my own corpus on colab gpu but I am getting an error of resource exhausted
can someone tell me how to fix this

Also what are the expected output of this further pretraining
Are they the bert tenserflow files that we can use for fine-tuning ( checkpoint, config, and vocab)?

Thank u

sorry for the late answer!
i am not very familiar with tensorflow, but there are some suggestions:

  1. check the version of tensorflow and make sure it is 1.1x
  2. if it has OOM problems, please reduce your batch size or reduce your max sequence length. the official bert repo has provided an example:
    image
  3. we do not have some resources for fine-tuning with tensorflow, you can check from the official bert repo if you want

from bert4doc-classification.

xuyige avatar xuyige commented on August 16, 2024

Hey men, I encounter the same issue. Are you able to resolve it?
I keep getting this,
OP_REQUIRES failed at cwise_ops_common.cc:70 : Resource exhausted: OOM when allocating tensor with shape[768,768].
I already reduce the batch size to 3 but didn't work.

sorry for the late answer!
if you have OOM problems, please reduce your batch size and max sequence length
the official bert repo has provided some example with a 12G GPU

from bert4doc-classification.

rajae-Bens avatar rajae-Bens commented on August 16, 2024

Hi,

thank u for answering

I reduced the train_batch_size to 8 and max_seq_length to 40

but I still get the resource exhausted error

I am running the code on colab gpu 12G RAM
any ideas plz

thank u

from bert4doc-classification.

xuyige avatar xuyige commented on August 16, 2024

Hi,

thank u for answering

I reduced the train_batch_size to 8 and max_seq_length to 40

but I still get the resource exhausted error

I am running the code on colab gpu 12G RAM
any ideas plz

thank u

as your description:
does your model contain some other NN modules?
does your colab gpu need to share with others?
do you have enough cpu sources (e.g., it is cpu OOM but not gpu OOM)?

from bert4doc-classification.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.