Spaces:
Runtime error
Runtime error
File size: 1,327 Bytes
2e82449 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 |
# from torch.utils.checkpoint import checkpoint
#
# import ldm.modules.attention
# import ldm.modules.diffusionmodules.openaimodel
#
#
# def BasicTransformerBlock_forward(self, x, context=None):
# return checkpoint(self._forward, x, context)
#
#
# def AttentionBlock_forward(self, x):
# return checkpoint(self._forward, x)
#
#
# def ResBlock_forward(self, x, emb):
# return checkpoint(self._forward, x, emb)
#
#
# stored = []
#
#
# def add():
# if len(stored) != 0:
# return
#
# stored.extend([
# ldm.modules.attention.BasicTransformerBlock.forward,
# ldm.modules.diffusionmodules.openaimodel.ResBlock.forward,
# ldm.modules.diffusionmodules.openaimodel.AttentionBlock.forward
# ])
#
# ldm.modules.attention.BasicTransformerBlock.forward = BasicTransformerBlock_forward
# ldm.modules.diffusionmodules.openaimodel.ResBlock.forward = ResBlock_forward
# ldm.modules.diffusionmodules.openaimodel.AttentionBlock.forward = AttentionBlock_forward
#
#
# def remove():
# if len(stored) == 0:
# return
#
# ldm.modules.attention.BasicTransformerBlock.forward = stored[0]
# ldm.modules.diffusionmodules.openaimodel.ResBlock.forward = stored[1]
# ldm.modules.diffusionmodules.openaimodel.AttentionBlock.forward = stored[2]
#
# stored.clear()
#
|