Update modeling_gigarembed.py
Browse files- modeling_gigarembed.py +0 -4
modeling_gigarembed.py
CHANGED
|
@@ -289,10 +289,6 @@ class LatentAttentionModel(PreTrainedModel):
|
|
| 289 |
FeedForward(latent_dim),
|
| 290 |
])
|
| 291 |
|
| 292 |
-
self.w_lexical = torch.nn.Linear(latent_dim, 1)
|
| 293 |
-
self.w_multi_vector = torch.nn.Linear(latent_dim, latent_dim)
|
| 294 |
-
|
| 295 |
-
# self.output_normalize = config.output_normalize
|
| 296 |
self.register_parameter("latents", torch.nn.Parameter(torch.randn(num_latents, latent_dim)))
|
| 297 |
self._attn_implementation = "eager"
|
| 298 |
|
|
|
|
| 289 |
FeedForward(latent_dim),
|
| 290 |
])
|
| 291 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 292 |
self.register_parameter("latents", torch.nn.Parameter(torch.randn(num_latents, latent_dim)))
|
| 293 |
self._attn_implementation = "eager"
|
| 294 |
|