Skip to content

Commit dc1d90d

Browse files
committed
Update package naming from Mambular to DeepTabular in docs/homepage.md
1 parent d41819c commit dc1d90d

1 file changed

Lines changed: 27 additions & 27 deletions

File tree

docs/homepage.md

Lines changed: 27 additions & 27 deletions
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,13 @@
1-
# Mambular: Tabular Deep Learning Made Simple
1+
# DeepTabular: Tabular Deep Learning Made Simple
22

3-
Mambular is a Python library for tabular deep learning. It includes models that leverage the Mamba (State Space Model) architecture, as well as other popular models like TabTransformer, FTTransformer, TabM and tabular ResNets. Check out our paper `Mambular: A Sequential Model for Tabular Deep Learning`, available [here](https://arxiv.org/abs/2408.06291). Also check out our paper introducing [TabulaRNN](https://arxiv.org/pdf/2411.17207) and analyzing the efficiency of NLP inspired tabular models.
3+
DeepTabular is a Python library for tabular deep learning. It includes models that leverage the Mamba (State Space Model) architecture, as well as other popular models like TabTransformer, FTTransformer, TabM and tabular ResNets. Check out our paper `Mambular: A Sequential Model for Tabular Deep Learning`, available [here](https://arxiv.org/abs/2408.06291). Also check out our paper introducing [TabulaRNN](https://arxiv.org/pdf/2411.17207) and analyzing the efficiency of NLP inspired tabular models.
44

55

66
# 🏃 Quickstart
7-
Similar to any sklearn model, Mambular models can be fit as easy as this:
7+
Similar to any sklearn model, DeepTabular models can be fit as easy as this:
88

99
```python
10-
from mambular.models import MambularClassifier
10+
from deeptabular.models import MambularClassifier
1111
# Initialize and fit your model
1212
model = MambularClassifier()
1313

@@ -16,7 +16,7 @@ model.fit(X, y, max_epochs=150, lr=1e-04)
1616
```
1717

1818
# 📖 Introduction
19-
Mambular is a Python package that brings the power of advanced deep learning architectures to tabular data, offering a suite of models for regression, classification, and distributional regression tasks. Designed with ease of use in mind, Mambular models adhere to scikit-learn's `BaseEstimator` interface, making them highly compatible with the familiar scikit-learn ecosystem. This means you can fit, predict, and evaluate using Mambular models just as you would with any traditional scikit-learn model, but with the added performance and flexibility of deep learning.
19+
DeepTabular is a Python package that brings the power of advanced deep learning architectures to tabular data, offering a suite of models for regression, classification, and distributional regression tasks. Designed with ease of use in mind, DeepTabular models adhere to scikit-learn's `BaseEstimator` interface, making them highly compatible with the familiar scikit-learn ecosystem. This means you can fit, predict, and evaluate using DeepTabular models just as you would with any traditional scikit-learn model, but with the added performance and flexibility of deep learning.
2020

2121

2222
# 🤖 Models
@@ -44,13 +44,13 @@ Hence, they are available as e.g. `MambularRegressor`, `MambularClassifier` or `
4444

4545
# 📚 Documentation
4646

47-
You can find the Mamba-Tabular API documentation [here](https://mambular.readthedocs.io/en/latest/).
47+
You can find the DeepTabular API documentation [here](https://deeptabular.readthedocs.io/en/latest/).
4848

4949
# 🛠️ Installation
5050

51-
Install Mambular using pip:
51+
Install DeepTabular using pip:
5252
```sh
53-
pip install mambular
53+
pip install deeptabular
5454
```
5555

5656
If you want to use the original mamba and mamba2 implementations, additionally install mamba-ssm via:
@@ -70,7 +70,7 @@ pip install mamba-ssm
7070

7171
<h2> Preprocessing </h2>
7272

73-
Mambular simplifies data preprocessing with a range of tools designed for easy transformation of tabular data.
73+
DeepTabular simplifies data preprocessing with a range of tools designed for easy transformation of tabular data.
7474

7575
<h3> Data Type Detection and Transformation </h3>
7676

@@ -90,10 +90,10 @@ Mambular simplifies data preprocessing with a range of tools designed for easy t
9090

9191
<h2> Fit a Model </h2>
9292

93-
Fitting a model in mambular is as simple as it gets. All models in mambular are sklearn BaseEstimators. Thus, the `fit` method is implemented for all of them. Additionally, this allows for using all other sklearn inherent methods such as their built in hyperparameter optimization tools.
93+
Fitting a model in deeptabular is as simple as it gets. All models in deeptabular are sklearn BaseEstimators. Thus, the `fit` method is implemented for all of them. Additionally, this allows for using all other sklearn inherent methods such as their built in hyperparameter optimization tools.
9494

9595
```python
96-
from mambular.models import MambularClassifier
96+
from deeptabular.models import MambularClassifier
9797
# Initialize and fit your model
9898
model = MambularClassifier(
9999
d_model=64,
@@ -171,12 +171,12 @@ Or use the built-in bayesian hpo simply by running:
171171
best_params = model.optimize_hparams(X, y)
172172
```
173173

174-
This automatically sets the search space based on the default config from ``mambular.configs``. See the documentation for all params with regard to ``optimize_hparams()``. However, the preprocessor arguments are fixed and cannot be optimized here.
174+
This automatically sets the search space based on the default config from ``deeptabular.configs``. See the documentation for all params with regard to ``optimize_hparams()``. However, the preprocessor arguments are fixed and cannot be optimized here.
175175

176176

177177
<h2> ⚖️ Distributional Regression with MambularLSS </h2>
178178

179-
MambularLSS allows you to model the full distribution of a response variable, not just its mean. This is crucial when understanding variability, skewness, or kurtosis is important. All Mambular models are available as distributional models.
179+
MambularLSS allows you to model the full distribution of a response variable, not just its mean. This is crucial when understanding variability, skewness, or kurtosis is important. All DeepTabular models are available as distributional models.
180180

181181
<h3> Key Features of MambularLSS: </h3>
182182

@@ -203,10 +203,10 @@ These distribution classes make MambularLSS versatile in modeling various data t
203203

204204
<h3> Getting Started with MambularLSS: </h3>
205205

206-
To integrate distributional regression into your workflow with `MambularLSS`, start by initializing the model with your desired configuration, similar to other Mambular models:
206+
To integrate distributional regression into your workflow with `MambularLSS`, start by initializing the model with your desired configuration, similar to other DeepTabular models:
207207

208208
```python
209-
from mambular.models import MambularLSS
209+
from deeptabular.models import MambularLSS
210210

211211
# Initialize the MambularLSS model
212212
model = MambularLSS(
@@ -231,11 +231,11 @@ model.fit(
231231

232232
# 💻 Implement Your Own Model
233233

234-
Mambular allows users to easily integrate their custom models into the existing logic. This process is designed to be straightforward, making it simple to create a PyTorch model and define its forward pass. Instead of inheriting from `nn.Module`, you inherit from Mambular's `BaseModel`. Each Mambular model takes three main arguments: the number of classes (e.g., 1 for regression or 2 for binary classification), `cat_feature_info`, and `num_feature_info` for categorical and numerical feature information, respectively. Additionally, you can provide a config argument, which can either be a custom configuration or one of the provided default configs.
234+
DeepTabular allows users to easily integrate their custom models into the existing logic. This process is designed to be straightforward, making it simple to create a PyTorch model and define its forward pass. Instead of inheriting from `nn.Module`, you inherit from DeepTabular's `BaseModel`. Each DeepTabular model takes three main arguments: the number of classes (e.g., 1 for regression or 2 for binary classification), `cat_feature_info`, and `num_feature_info` for categorical and numerical feature information, respectively. Additionally, you can provide a config argument, which can either be a custom configuration or one of the provided default configs.
235235

236-
One of the key advantages of using Mambular is that the inputs to the forward passes are lists of tensors. While this might be unconventional, it is highly beneficial for models that treat different data types differently. For example, the TabTransformer model leverages this feature to handle categorical and numerical data separately, applying different transformations and processing steps to each type of data.
236+
One of the key advantages of using DeepTabular is that the inputs to the forward passes are lists of tensors. While this might be unconventional, it is highly beneficial for models that treat different data types differently. For example, the TabTransformer model leverages this feature to handle categorical and numerical data separately, applying different transformations and processing steps to each type of data.
237237

238-
Here's how you can implement a custom model with Mambular:
238+
Here's how you can implement a custom model with DeepTabular:
239239

240240
1. **First, define your config:**
241241
The configuration class allows you to specify hyperparameters and other settings for your model. This can be done using a simple dataclass.
@@ -255,8 +255,8 @@ Here's how you can implement a custom model with Mambular:
255255
Define your custom model just as you would for an `nn.Module`. The main difference is that you will inherit from `BaseModel` and use the provided feature information to construct your layers. To integrate your model into the existing API, you only need to define the architecture and the forward pass.
256256

257257
```python
258-
from mambular.base_models import BaseModel
259-
from mambular.utils.get_feature_dimensions import get_feature_dimensions
258+
from deeptabular.base_models import BaseModel
259+
from deeptabular.utils.get_feature_dimensions import get_feature_dimensions
260260
import torch
261261
import torch.nn
262262

@@ -285,19 +285,19 @@ Here's how you can implement a custom model with Mambular:
285285
return output
286286
```
287287

288-
3. **Leverage the Mambular API:**
289-
You can build a regression, classification, or distributional regression model that can leverage all of Mambular's built-in methods by using the following:
288+
3. **Leverage the DeepTabular API:**
289+
You can build a regression, classification, or distributional regression model that can leverage all of DeepTabular's built-in methods by using the following:
290290

291291
```python
292-
from mambular.models import SklearnBaseRegressor
292+
from deeptabular.models import SklearnBaseRegressor
293293

294294
class MyRegressor(SklearnBaseRegressor):
295295
def __init__(self, **kwargs):
296296
super().__init__(model=MyCustomModel, config=MyConfig, **kwargs)
297297
```
298298

299299
4. **Train and evaluate your model:**
300-
You can now fit, evaluate, and predict with your custom model just like with any other Mambular model. For classification or distributional regression, inherit from `SklearnBaseClassifier` or `SklearnBaseLSS` respectively.
300+
You can now fit, evaluate, and predict with your custom model just like with any other DeepTabular model. For classification or distributional regression, inherit from `SklearnBaseClassifier` or `SklearnBaseLSS` respectively.
301301

302302
```python
303303
regressor = MyRegressor(numerical_preprocessing="ple")
@@ -306,15 +306,15 @@ Here's how you can implement a custom model with Mambular:
306306

307307
# Custom Training
308308

309-
If you prefer to setup custom training, preprocessing and evaluation, you can simply use the `mambular.base_models`.
309+
If you prefer to setup custom training, preprocessing and evaluation, you can simply use the `deeptabular.base_models`.
310310
Just be careful that all basemodels expect lists of features as inputs. More precisely as list for numerical features and a list for categorical features. A custom training loop, with random data could look like this.
311311

312312
```python
313313
import torch
314314
import torch.nn as nn
315315
import torch.optim as optim
316-
from mambular.base_models import Mambular
317-
from mambular.configs import DefaultMambularConfig
316+
from deeptabular.base_models import Mambular
317+
from deeptabular.configs import DefaultMambularConfig
318318

319319
# Dummy data and configuration
320320
cat_feature_info = {

0 commit comments

Comments
 (0)