提交 5d03f95e 编写于 作者: J jrzaurin

updated to new version of black and added quite to isort to avoid simlink warnings

上级 905e4d75
# sort imports # sort imports
isort . pytorch_widedeep tests examples setup.py isort --quiet . pytorch_widedeep tests examples setup.py
# Black code style # Black code style
black . pytorch_widedeep tests examples setup.py black . pytorch_widedeep tests examples setup.py
# flake8 standards # flake8 standards
......
...@@ -104,4 +104,6 @@ if __name__ == "__main__": ...@@ -104,4 +104,6 @@ if __name__ == "__main__":
# model = WideDeep(wide=wide, deepdense=deepdense) # model = WideDeep(wide=wide, deepdense=deepdense)
# model.load_state_dict(torch.load("model_weights/model_dict.t")) # model.load_state_dict(torch.load("model_weights/model_dict.t"))
# # <All keys matched successfully> # # <All keys matched successfully>
import pdb; pdb.set_trace() # breakpoint dde47114 // import pdb
pdb.set_trace() # breakpoint dde47114 //
...@@ -166,8 +166,7 @@ class DeepImage(nn.Module): ...@@ -166,8 +166,7 @@ class DeepImage(nn.Module):
self.output_dim = head_layers[-1] self.output_dim = head_layers[-1]
def forward(self, x: Tensor) -> Tensor: # type: ignore def forward(self, x: Tensor) -> Tensor: # type: ignore
r"""Forward pass connecting the `'backbone'` with the `'head layers'` r"""Forward pass connecting the `'backbone'` with the `'head layers'`"""
"""
x = self.backbone(x) x = self.backbone(x)
x = x.view(x.size(0), -1) x = x.view(x.size(0), -1)
if self.head_layers is not None: if self.head_layers is not None:
......
...@@ -701,7 +701,7 @@ class WideDeep(nn.Module): ...@@ -701,7 +701,7 @@ class WideDeep(nn.Module):
X_test: Optional[Dict[str, np.ndarray]] = None, X_test: Optional[Dict[str, np.ndarray]] = None,
) -> np.ndarray: ) -> np.ndarray:
r"""Returns the predicted probabilities for the test dataset for binary r"""Returns the predicted probabilities for the test dataset for binary
and multiclass methods and multiclass methods
""" """
preds_l = self._predict(X_wide, X_deep, X_text, X_img, X_test) preds_l = self._predict(X_wide, X_deep, X_text, X_img, X_test)
if self.method == "binary": if self.method == "binary":
......
...@@ -93,15 +93,16 @@ class WidePreprocessor(BasePreprocessor): ...@@ -93,15 +93,16 @@ class WidePreprocessor(BasePreprocessor):
""" """
def __init__( def __init__(
self, wide_cols: List[str], crossed_cols=None, self,
wide_cols: List[str],
crossed_cols=None,
): ):
super(WidePreprocessor, self).__init__() super(WidePreprocessor, self).__init__()
self.wide_cols = wide_cols self.wide_cols = wide_cols
self.crossed_cols = crossed_cols self.crossed_cols = crossed_cols
def fit(self, df: pd.DataFrame) -> BasePreprocessor: def fit(self, df: pd.DataFrame) -> BasePreprocessor:
"""Fits the Preprocessor and creates required attributes """Fits the Preprocessor and creates required attributes"""
"""
df_wide = self._prepare_wide(df) df_wide = self._prepare_wide(df)
self.wide_crossed_cols = df_wide.columns.tolist() self.wide_crossed_cols = df_wide.columns.tolist()
vocab = self._make_global_feature_list(df_wide[self.wide_crossed_cols]) vocab = self._make_global_feature_list(df_wide[self.wide_crossed_cols])
...@@ -110,8 +111,7 @@ class WidePreprocessor(BasePreprocessor): ...@@ -110,8 +111,7 @@ class WidePreprocessor(BasePreprocessor):
return self return self
def transform(self, df: pd.DataFrame) -> np.array: def transform(self, df: pd.DataFrame) -> np.array:
r"""Returns the processed dataframe r"""Returns the processed dataframe"""
"""
try: try:
self.feature_dict self.feature_dict
except: except:
...@@ -147,8 +147,7 @@ class WidePreprocessor(BasePreprocessor): ...@@ -147,8 +147,7 @@ class WidePreprocessor(BasePreprocessor):
return decoded return decoded
def fit_transform(self, df: pd.DataFrame) -> np.ndarray: def fit_transform(self, df: pd.DataFrame) -> np.ndarray:
"""Combines ``fit`` and ``transform`` """Combines ``fit`` and ``transform``"""
"""
return self.fit(df).transform(df) return self.fit(df).transform(df)
def _make_global_feature_list(self, df: pd.DataFrame) -> List: def _make_global_feature_list(self, df: pd.DataFrame) -> List:
...@@ -256,8 +255,7 @@ class DensePreprocessor(BasePreprocessor): ...@@ -256,8 +255,7 @@ class DensePreprocessor(BasePreprocessor):
), "'embed_cols' and 'continuous_cols' are 'None'. Please, define at least one of the two." ), "'embed_cols' and 'continuous_cols' are 'None'. Please, define at least one of the two."
def fit(self, df: pd.DataFrame) -> BasePreprocessor: def fit(self, df: pd.DataFrame) -> BasePreprocessor:
"""Fits the Preprocessor and creates required attributes """Fits the Preprocessor and creates required attributes"""
"""
if self.embed_cols is not None: if self.embed_cols is not None:
df_emb = self._prepare_embed(df) df_emb = self._prepare_embed(df)
self.label_encoder = LabelEncoder(df_emb.columns.tolist()).fit(df_emb) self.label_encoder = LabelEncoder(df_emb.columns.tolist()).fit(df_emb)
...@@ -274,8 +272,7 @@ class DensePreprocessor(BasePreprocessor): ...@@ -274,8 +272,7 @@ class DensePreprocessor(BasePreprocessor):
return self return self
def transform(self, df: pd.DataFrame) -> np.ndarray: def transform(self, df: pd.DataFrame) -> np.ndarray:
"""Returns the processed ``dataframe`` as a np.ndarray """Returns the processed ``dataframe`` as a np.ndarray"""
"""
if self.embed_cols is not None: if self.embed_cols is not None:
df_emb = self._prepare_embed(df) df_emb = self._prepare_embed(df)
df_emb = self.label_encoder.transform(df_emb) df_emb = self.label_encoder.transform(df_emb)
...@@ -302,8 +299,7 @@ class DensePreprocessor(BasePreprocessor): ...@@ -302,8 +299,7 @@ class DensePreprocessor(BasePreprocessor):
return df_deep.values return df_deep.values
def fit_transform(self, df: pd.DataFrame) -> np.ndarray: def fit_transform(self, df: pd.DataFrame) -> np.ndarray:
"""Combines ``fit`` and ``transform`` """Combines ``fit`` and ``transform``"""
"""
return self.fit(df).transform(df) return self.fit(df).transform(df)
def _prepare_embed(self, df: pd.DataFrame) -> pd.DataFrame: def _prepare_embed(self, df: pd.DataFrame) -> pd.DataFrame:
...@@ -387,8 +383,7 @@ class TextPreprocessor(BasePreprocessor): ...@@ -387,8 +383,7 @@ class TextPreprocessor(BasePreprocessor):
self.verbose = verbose self.verbose = verbose
def fit(self, df: pd.DataFrame) -> BasePreprocessor: def fit(self, df: pd.DataFrame) -> BasePreprocessor:
"""Builds the vocabulary """Builds the vocabulary"""
"""
texts = df[self.text_col].tolist() texts = df[self.text_col].tolist()
tokens = get_texts(texts) tokens = get_texts(texts)
self.vocab = Vocab.create( self.vocab = Vocab.create(
...@@ -399,8 +394,7 @@ class TextPreprocessor(BasePreprocessor): ...@@ -399,8 +394,7 @@ class TextPreprocessor(BasePreprocessor):
return self return self
def transform(self, df: pd.DataFrame) -> np.ndarray: def transform(self, df: pd.DataFrame) -> np.ndarray:
"""Returns the padded, `numericalised` sequences """Returns the padded, `numericalised` sequences"""
"""
try: try:
self.vocab self.vocab
except: except:
...@@ -419,8 +413,7 @@ class TextPreprocessor(BasePreprocessor): ...@@ -419,8 +413,7 @@ class TextPreprocessor(BasePreprocessor):
return padded_seq return padded_seq
def fit_transform(self, df: pd.DataFrame) -> np.ndarray: def fit_transform(self, df: pd.DataFrame) -> np.ndarray:
"""Combines ``fit`` and ``transform`` """Combines ``fit`` and ``transform``"""
"""
return self.fit(df).transform(df) return self.fit(df).transform(df)
...@@ -502,8 +495,7 @@ class ImagePreprocessor(BasePreprocessor): ...@@ -502,8 +495,7 @@ class ImagePreprocessor(BasePreprocessor):
return self return self
def transform(self, df: pd.DataFrame) -> np.ndarray: def transform(self, df: pd.DataFrame) -> np.ndarray:
"""Resizes the images to the input height and width. """Resizes the images to the input height and width."""
"""
try: try:
self.aap self.aap
except: except:
...@@ -564,6 +556,5 @@ class ImagePreprocessor(BasePreprocessor): ...@@ -564,6 +556,5 @@ class ImagePreprocessor(BasePreprocessor):
return np.asarray(resized_imgs) return np.asarray(resized_imgs)
def fit_transform(self, df: pd.DataFrame) -> np.ndarray: def fit_transform(self, df: pd.DataFrame) -> np.ndarray:
"""Combines ``fit`` and ``transform`` """Combines ``fit`` and ``transform``"""
"""
return self.fit(df).transform(df) return self.fit(df).transform(df)
...@@ -45,8 +45,7 @@ class LabelEncoder(object): ...@@ -45,8 +45,7 @@ class LabelEncoder(object):
self.columns_to_encode = columns_to_encode self.columns_to_encode = columns_to_encode
def fit(self, df: pd.DataFrame) -> "LabelEncoder": def fit(self, df: pd.DataFrame) -> "LabelEncoder":
"""Creates encoding attributes """Creates encoding attributes"""
"""
df_inp = df.copy() df_inp = df.copy()
...@@ -78,8 +77,7 @@ class LabelEncoder(object): ...@@ -78,8 +77,7 @@ class LabelEncoder(object):
return self return self
def transform(self, df: pd.DataFrame) -> pd.DataFrame: def transform(self, df: pd.DataFrame) -> pd.DataFrame:
"""Label Encoded the categories in ``columns_to_encode`` """Label Encoded the categories in ``columns_to_encode``"""
"""
try: try:
self.encoding_dict self.encoding_dict
except AttributeError: except AttributeError:
...@@ -126,8 +124,7 @@ class LabelEncoder(object): ...@@ -126,8 +124,7 @@ class LabelEncoder(object):
return self.fit(df).transform(df) return self.fit(df).transform(df)
def inverse_transform(self, df: pd.DataFrame) -> pd.DataFrame: def inverse_transform(self, df: pd.DataFrame) -> pd.DataFrame:
"""Returns the original categories """Returns the original categories"""
"""
for k, v in self.inverse_encoding_dict.items(): for k, v in self.inverse_encoding_dict.items():
df[k] = df[k].apply(lambda x: v[x]) df[k] = df[k].apply(lambda x: v[x])
......
...@@ -78,8 +78,7 @@ defaults.text_spec_tok = [UNK, PAD, BOS, EOS, FLD, TK_MAJ, TK_UP, TK_REP, TK_WRE ...@@ -78,8 +78,7 @@ defaults.text_spec_tok = [UNK, PAD, BOS, EOS, FLD, TK_MAJ, TK_UP, TK_REP, TK_WRE
class BaseTokenizer: class BaseTokenizer:
"""Basic class for a tokenizer function. """Basic class for a tokenizer function."""
"""
def __init__(self, lang: str): def __init__(self, lang: str):
self.lang = lang self.lang = lang
...@@ -278,8 +277,7 @@ class Tokenizer: ...@@ -278,8 +277,7 @@ class Tokenizer:
return toks return toks
def _process_all_1(self, texts: Collection[str]) -> List[List[str]]: def _process_all_1(self, texts: Collection[str]) -> List[List[str]]:
"""Process a list of ``texts`` in one process. """Process a list of ``texts`` in one process."""
"""
tok = self.tok_func(self.lang) tok = self.tok_func(self.lang)
if self.special_cases: if self.special_cases:
...@@ -332,13 +330,11 @@ class Vocab: ...@@ -332,13 +330,11 @@ class Vocab:
self.stoi = defaultdict(int, {v: k for k, v in enumerate(self.itos)}) self.stoi = defaultdict(int, {v: k for k, v in enumerate(self.itos)})
def numericalize(self, t: Collection[str]) -> List[int]: def numericalize(self, t: Collection[str]) -> List[int]:
"""Convert a list of str (or tokens) ``t`` to their ids. """Convert a list of str (or tokens) ``t`` to their ids."""
"""
return [self.stoi[w] for w in t] return [self.stoi[w] for w in t]
def textify(self, nums: Collection[int], sep=" ") -> List[str]: def textify(self, nums: Collection[int], sep=" ") -> List[str]:
"""Convert a list of ``nums`` (or indexes) to their tokens. """Convert a list of ``nums`` (or indexes) to their tokens."""
"""
return sep.join([self.itos[i] for i in nums]) if sep is not None else [self.itos[i] for i in nums] # type: ignore return sep.join([self.itos[i] for i in nums]) if sep is not None else [self.itos[i] for i in nums] # type: ignore
def __getstate__(self): def __getstate__(self):
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册