Skip to content

Gradient Explode #33

@HsiaoYingLu

Description

@HsiaoYingLu

Hi all,

I wonder if anyone encountered gradient explosion after training for a few epochs?
I am trying to use the CAN losses on a sequence generation model.
But, the gradient of generator just kept becoming too large for that the sequence generated can still be classified too well.
For now, I kept training even if the gradients are too large every 2 epochs and I don't see the result of my model being anywhere near creative.
Any thought on this are appreciated!

Thank you

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions