Skip to content
This repository was archived by the owner on Aug 28, 2025. It is now read-only.

Commit 9334a3e

Browse files
[pre-commit.ci] pre-commit suggestions (#279)
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com> Co-authored-by: Jirka Borovec <6035284+Borda@users.noreply.github.com>
1 parent 57e7fe6 commit 9334a3e

7 files changed

Lines changed: 12 additions & 13 deletions

File tree

.actions/assistant.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -188,7 +188,7 @@ def _load_meta(folder: str, strict: bool = False) -> Optional[dict]:
188188
189189
Args:
190190
folder: path to the folder with python script, meta and artefacts
191-
strict: raise error if meta is missing required feilds
191+
strict: raise error if meta is missing required fields
192192
"""
193193
fpath = AssistantCLI._find_meta(folder)
194194
assert fpath, f"Missing meta file in folder: {folder}"

.pre-commit-config.yaml

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -23,21 +23,21 @@ repos:
2323
- id: detect-private-key
2424

2525
- repo: https://github.com/asottile/pyupgrade
26-
rev: v3.8.0
26+
rev: v3.14.0
2727
hooks:
2828
- id: pyupgrade
2929
args: ["--py38-plus"]
3030
name: Upgrade code
3131

3232
- repo: https://github.com/codespell-project/codespell
33-
rev: v2.2.5
33+
rev: v2.2.6
3434
hooks:
3535
- id: codespell
3636
additional_dependencies: [tomli]
3737
#args: ["--write-changes"]
3838

3939
- repo: https://github.com/PyCQA/docformatter
40-
rev: v1.7.3
40+
rev: v1.7.5
4141
hooks:
4242
- id: docformatter
4343
args:
@@ -53,7 +53,7 @@ repos:
5353
args: ["--print-width=120"]
5454

5555
- repo: https://github.com/psf/black
56-
rev: 23.3.0
56+
rev: 23.9.1
5757
hooks:
5858
- id: black
5959
name: Format code
@@ -64,7 +64,7 @@ repos:
6464
- id: yesqa
6565

6666
- repo: https://github.com/executablebooks/mdformat
67-
rev: 0.7.16
67+
rev: 0.7.17
6868
hooks:
6969
- id: mdformat
7070
additional_dependencies:
@@ -73,7 +73,7 @@ repos:
7373
- mdformat_frontmatter
7474

7575
- repo: https://github.com/astral-sh/ruff-pre-commit
76-
rev: v0.0.276
76+
rev: v0.0.292
7777
hooks:
7878
- id: ruff
7979
args: ["--fix"]

course_UvA-DL/06-graph-neural-networks/GNN_overview.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -843,7 +843,7 @@ def print_results(result_dict):
843843
# In this case, we will use the average pooling.
844844
# Hence, we need to know which nodes should be included in which average pool.
845845
# Using this pooling, we can already create our graph network below.
846-
# Specifically, we re-use our class `GNNModel` from before,
846+
# Specifically, we reuse our class `GNNModel` from before,
847847
# and simply add an average pool and single linear layer for the graph prediction task.
848848

849849

course_UvA-DL/07-deep-energy-based-generative-models/Deep_Energy_Models.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -315,7 +315,7 @@ def forward(self, x):
315315
# inside the MCMC sampling to obtain reasonable samples.
316316
# However, there is a training trick that significantly reduces the sampling cost: using a sampling buffer.
317317
# The idea is that we store the samples of the last couple of batches in a buffer,
318-
# and re-use those as the starting point of the MCMC algorithm for the next batches.
318+
# and reuse those as the starting point of the MCMC algorithm for the next batches.
319319
# This reduces the sampling cost because the model requires a significantly
320320
# lower number of steps to converge to reasonable samples.
321321
# However, to not solely rely on previous samples and allow novel samples as well,

course_UvA-DL/10-autoregressive-image-modeling/Autoregressive_Image_Modeling.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -905,7 +905,7 @@ def autocomplete_image(img):
905905
# potentially undesirable behavior. For instance, the value 242 has a
906906
# 1000x lower likelihood than 243 although they are extremely close and
907907
# can often not be distinguished. This shows that the model might have not
908-
# generlized well over pixel values. The better solution to this problem
908+
# generalized well over pixel values. The better solution to this problem
909909
# is to use discrete logitics mixtures instead of a softmax distribution.
910910
# A discrete logistic distribution can be imagined as discretized, binned
911911
# Gaussians. Using a mixture of discrete logistics instead of a softmax

flash_tutorials/text_classification/text_classification.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
#
66
# Finetuning consists of four steps:
77
#
8-
# - 1. Train a source neural network model on a source dataset. For text classication, it is traditionally a transformer model such as BERT [Bidirectional Encoder Representations from Transformers](https://arxiv.org/abs/1810.04805) trained on wikipedia.
8+
# - 1. Train a source neural network model on a source dataset. For text classification, it is traditionally a transformer model such as BERT [Bidirectional Encoder Representations from Transformers](https://arxiv.org/abs/1810.04805) trained on wikipedia.
99
# As those model are costly to train, [Transformers](https://github.com/huggingface/transformers) or [FairSeq](https://github.com/pytorch/fairseq) libraries provides popular pre-trained model architectures for NLP. In this notebook, we will be using [tiny-bert](https://huggingface.co/prajjwal1/bert-tiny).
1010
#
1111
# - 2. Create a new neural network the target model. Its architecture replicates all model designs and their parameters on the source model, expect the latest layer which is removed. This model without its latest layers is traditionally called a backbone

lightning_examples/finetuning-scheduler/finetuning-scheduler.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -235,8 +235,7 @@ def __init__(
235235
tokenizers_parallelism: bool = True,
236236
**dataloader_kwargs: Any,
237237
):
238-
r"""Initialize the ``LightningDataModule`` designed for both the RTE or BoolQ SuperGLUE Hugging Face
239-
datasets.
238+
r"""Initialize the ``LightningDataModule`` designed for both the RTE or BoolQ SuperGLUE Hugging Face datasets.
240239
241240
Args:
242241
model_name_or_path (str):

0 commit comments

Comments
 (0)