提交 8cd49444 编写于 作者: J jrzaurin

Further refined README

上级 96dc01b6
...@@ -22,7 +22,7 @@ using wide and deep models. ...@@ -22,7 +22,7 @@ using wide and deep models.
**Experiments and comparisson with `LightGBM`**: [TabularDL vs LightGBM](https://github.com/jrzaurin/tabulardl-benchmark) **Experiments and comparisson with `LightGBM`**: [TabularDL vs LightGBM](https://github.com/jrzaurin/tabulardl-benchmark)
**slack**: if you want to contribute or just want to chat with us, join [slack](https://join.slack.com/t/pytorch-widedeep/shared_invite/zt-soss7stf-iXpVuLeKZz8lGTnxxtHtTw) **Slack**: if you want to contribute or just want to chat with us, join [slack](https://join.slack.com/t/pytorch-widedeep/shared_invite/zt-soss7stf-iXpVuLeKZz8lGTnxxtHtTw)
### Introduction ### Introduction
...@@ -109,14 +109,14 @@ is an adaptation of the original implementation. ...@@ -109,14 +109,14 @@ is an adaptation of the original implementation.
5. ``FT-Transformer``: or Feature Tokenizer transformer. This is a relatively small 5. ``FT-Transformer``: or Feature Tokenizer transformer. This is a relatively small
variation of the ``TabTransformer``. The variation itself was first variation of the ``TabTransformer``. The variation itself was first
introduced in the ``SAINT`` paper, but the name ``FT-Transformer`` was first introduced in the ``SAINT`` paper, but the name "``FT-Transformer``" was first
used in used in
[Revisiting Deep Learning Models for TabularData](https://arxiv.org/abs/2106.11959). [Revisiting Deep Learning Models for Tabular Data](https://arxiv.org/abs/2106.11959).
When using the ``FT-Transformer`` each continuous feature is "embedded" When using the ``FT-Transformer`` each continuous feature is "embedded"
(i.e. each one going through a 1-layer MLP with or without activation (i.e. going through a 1-layer MLP with or without activation function) and
function) and then passed through the attention blocks along with the then passed through the attention blocks along with the categorical features.
categorical features. This is available in ``pytorch-widedeep``'s This is available in ``pytorch-widedeep``'s ``TabTransformer`` by setting the
``TabTransformer`` by setting the parameter ``embed_continuous = True``. parameter ``embed_continuous = True``.
6. ``SAINT``: Details on SAINT can be found in: 6. ``SAINT``: Details on SAINT can be found in:
...@@ -161,20 +161,19 @@ cd pytorch-widedeep ...@@ -161,20 +161,19 @@ cd pytorch-widedeep
pip install -e . pip install -e .
``` ```
**Important note for Mac users**: at the time of writing (June-2021) the **Important note for Mac users**: at the time of writing the latest `torch`
latest `torch` release is `1.9`. Some past release is `1.9`. Some past [issues](https://stackoverflow.com/questions/64772335/pytorch-w-parallelnative-cpp206)
[issues](https://stackoverflow.com/questions/64772335/pytorch-w-parallelnative-cpp206) when running on Mac, present in previous versions, persist on this release
when running on Mac, present in previous versions, persist on this release and and the data-loaders will not run in parallel. In addition, since `python
the data-loaders will not run in parallel. In addition, since `python 3.8`, 3.8`, [the `multiprocessing` library start method changed from `'fork'` to`'spawn'`](https://docs.python.org/3/library/multiprocessing.html#contexts-and-start-methods).
[the `multiprocessing` library start method changed from `'fork'` to
`'spawn'`](https://docs.python.org/3/library/multiprocessing.html#contexts-and-start-methods).
This also affects the data-loaders (for any `torch` version) and they will This also affects the data-loaders (for any `torch` version) and they will
not run in parallel. Therefore, for Mac users I recommend using `python 3.6` not run in parallel. Therefore, for Mac users I recommend using `python
or `3.7` and `torch <= 1.6` (with the corresponding, consistent version of 3.6` or `3.7` and `torch <= 1.6` (with the corresponding, consistent
`torchvision`, e.g. `0.7.0` for `torch 1.6`). I do not want to force this version of `torchvision`, e.g. `0.7.0` for `torch 1.6`). I do not want to
versioning in the `setup.py` file since I expect that all these issues are force this versioning in the `setup.py` file since I expect that all these
fixed in the future. Therefore, after installing `pytorch-widedeep` via pip issues are fixed in the future. Therefore, after installing
or directly from github, downgrade `torch` and `torchvision` manually: `pytorch-widedeep` via pip or directly from github, downgrade `torch` and
`torchvision` manually:
```bash ```bash
pip install pytorch-widedeep pip install pytorch-widedeep
......
...@@ -17,7 +17,7 @@ using wide and deep models. ...@@ -17,7 +17,7 @@ using wide and deep models.
**Experiments and comparisson with `LightGBM`**: [TabularDL vs LightGBM](https://github.com/jrzaurin/tabulardl-benchmark) **Experiments and comparisson with `LightGBM`**: [TabularDL vs LightGBM](https://github.com/jrzaurin/tabulardl-benchmark)
**slack**: if you want to contribute or just want to chat with us, join [slack](https://join.slack.com/t/pytorch-widedeep/shared_invite/zt-soss7stf-iXpVuLeKZz8lGTnxxtHtTw) **Slack**: if you want to contribute or just want to chat with us, join [slack](https://join.slack.com/t/pytorch-widedeep/shared_invite/zt-soss7stf-iXpVuLeKZz8lGTnxxtHtTw)
### Introduction ### Introduction
...@@ -57,20 +57,20 @@ cd pytorch-widedeep ...@@ -57,20 +57,20 @@ cd pytorch-widedeep
pip install -e . pip install -e .
``` ```
**Important note for Mac users**: at the time of writing (June-2021) the **Important note for Mac users**: at the time of writing the latest `torch`
latest `torch` release is `1.9`. Some past release is `1.9`. Some past [issues](https://stackoverflow.com/questions/64772335/pytorch-w-parallelnative-cpp206)
[issues](https://stackoverflow.com/questions/64772335/pytorch-w-parallelnative-cpp206) when running on Mac, present in previous versions, persist on this release
when running on Mac, present in previous versions, persist on this release and and the data-loaders will not run in parallel. In addition, since `python
the data-loaders will not run in parallel. In addition, since `python 3.8`, 3.8`, [the `multiprocessing` library start method changed from `'fork'` to`'spawn'`](https://docs.python.org/3/library/multiprocessing.html#contexts-and-start-methods).
[the `multiprocessing` library start method changed from `'fork'` to
`'spawn'`](https://docs.python.org/3/library/multiprocessing.html#contexts-and-start-methods).
This also affects the data-loaders (for any `torch` version) and they will This also affects the data-loaders (for any `torch` version) and they will
not run in parallel. Therefore, for Mac users I recommend using `python 3.6` not run in parallel. Therefore, for Mac users I recommend using `python
or `3.7` and `torch <= 1.6` (with the corresponding, consistent version of 3.6` or `3.7` and `torch <= 1.6` (with the corresponding, consistent
`torchvision`, e.g. `0.7.0` for `torch 1.6`). I do not want to force this version of `torchvision`, e.g. `0.7.0` for `torch 1.6`). I do not want to
versioning in the `setup.py` file since I expect that all these issues are force this versioning in the `setup.py` file since I expect that all these
fixed in the future. Therefore, after installing `pytorch-widedeep` via pip issues are fixed in the future. Therefore, after installing
or directly from github, downgrade `torch` and `torchvision` manually: `pytorch-widedeep` via pip or directly from github, downgrade `torch` and
`torchvision` manually:
```bash ```bash
pip install pytorch-widedeep pip install pytorch-widedeep
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册