From 3edaeebce851bc6f1b9b8170b03637289ecba1a6 Mon Sep 17 00:00:00 2001 From: "pre-commit-ci[bot]" <66853113+pre-commit-ci[bot]@users.noreply.github.com> Date: Sat, 1 Mar 2025 09:07:04 +0000 Subject: [PATCH] [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci --- .../models/multimodal/site_encoders/encoders.py | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/pvnet/models/multimodal/site_encoders/encoders.py b/pvnet/models/multimodal/site_encoders/encoders.py index 7acb71ce..22003ef1 100644 --- a/pvnet/models/multimodal/site_encoders/encoders.py +++ b/pvnet/models/multimodal/site_encoders/encoders.py @@ -74,24 +74,24 @@ def __init__( ) def _calculate_attention(self, x): - i=0 + i = 0 for key in x.keys(): - if(i==17): - ans=key + if i == 17: + ans = key print(key, type(key)) - i+=1 + i += 1 gsp_ids = x[ans].squeeze().int() attention = self._attention_network(gsp_ids) return attention def _encode_value(self, x): # Shape: [batch size, sequence length, PV site] - i=0 + i = 0 for key in x.keys(): - if(i==6): - ans=key + if i == 6: + ans = key print(key, type(key)) - i+=1 + i += 1 pv_site_seqs = x[ans].float() batch_size = pv_site_seqs.shape[0]