Dicee

Latest version: v0.1.3.2

Safety actively analyzes 640400 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 2 of 2

1.0.5

Features
1. An [AbstractCallback](https://github.com/dice-group/dice-embeddings/blob/b891508412b318ea50f1af809462506ef7ed667c/core/abstracts.py#L591) class is implemented. Few callback classes are created to print info related to training, to save the model paramaters and apply polyak parameter ensemble model [PrintCallback](https://github.com/dice-group/dice-embeddings/blob/b891508412b318ea50f1af809462506ef7ed667c/core/callbacks.py#L35), [KGESaveCallback](https://github.com/dice-group/dice-embeddings/blob/b891508412b318ea50f1af809462506ef7ed667c/core/callbacks.py#L65), [PPE](https://github.com/dice-group/dice-embeddings/blob/b891508412b318ea50f1af809462506ef7ed667c/core/callbacks.py#L149).
2. Pandas, Modin and Polars can be used as a [backend](https://github.com/dice-group/dice-embeddings/tree/b891508412b318ea50f1af809462506ef7ed667c/core/read_preprocess_save_load_kg). Reading, preprocessing, saving and loading can be done in a parallel fashion.
3. [AccumulateEpochLossCallback](https://github.com/dice-group/dice-embeddings/blob/b891508412b318ea50f1af809462506ef7ed667c/core/callbacks.py#L12)
5. Gradient Accumulation is implemented.

Applications
1. A [function](https://github.com/dice-group/dice-embeddings/blob/main/core/knowledge_graph_embeddings.py#L26) for predicting conjunctive queries over knowledge graph is implemented.
2. A [function](https://github.com/dice-group/dice-embeddings/blob/b891508412b318ea50f1af809462506ef7ed667c/core/knowledge_graph_embeddings.py#L92) to detect missing triples is implemented.

1.0.4

Features
1. KvsSample technique implemented. KvsSample is KvsAll with selected tail entities. This technique reduces the memory usage during training as we can select the number of tail entities.
2. [Sharded Training tested](https://pytorch-lightning.readthedocs.io/en/stable/advanced/model_parallel.html#sharded-training)

Maintenance
1. Use Python 3.9
2. More tests are added
3. ReadMe is structured

Todos for the next release
1. Explicit Class Kronecker Decomposition at retriving embeddings

1.0.3

Features
Self-supervised Learning module: Pseudo-Labelling and [Conformal Credal Self-Supervised Learning](https://arxiv.org/abs/2205.15239) implemented.

Maintenance
1. Documentation & Instrations are improved.
2. Use Python 3.10 due to [PEP 635](https://peps.python.org/pep-0635/)

Todos for the next release
2. Consider using [Weights & Biases](https://wandb.ai/site)
3. Study [[Raymond Hettinger](https://twitter.com/raymondh)](https://twitter.com/raymondh/status/1533369936739016705) 's talk about Structural Pattern Matching in the Real World: New tooling, real code, problems solved.
4. Explicit Class Kronecker Decomposition at retriving embeddings

v1.0.2

Features

Batch Relaxation training strategy started.
Seed selection for the computation is available.
Input KG size reduction: Entities that do not occur X times can be removed.
Lower memory usage through selecting most efficient index type.
swifter is included to do dataframe().apply() via using all CPUs
_QMult with 11.4 B on DBpedia is succesfuly trained and deployed._

Maintenance

The title of the repo. has been changed.
Repo name has been changed.
Testing with three pytest setting is documented
Regression tests are extended.
More functions and classes are documented.

Todos for the next release
1. Use Python 3.10 to use [PEP 635](https://peps.python.org/pep-0635/)
2. Use Python 3.10 to benefit [from 10% performance increase](https://mail.python.org/pipermail/python-dev/2016-January/142945.html) and https://bugs.python.org/issue42093 <3
3. Consider using [Weights & Biases](https://wandb.ai/site)

1.0.1

**Features**

1. 1vsAll, KvsAll, Negative Sampling strategy available.
2. Continues training (training pre-trained model) is implemented
3. Deployment and opensourcing a pre-trained KG is implemented.
4. Using multi-CPUs during pre-processing is available. This is important to scale on large KGs
5. Artificial noise can be added into input KG. The rate of added noisy triples can be used to reguirlizer the selected KG, as well as, compare link prediction performances of KGE models under noisy data.

**Maintenance**
1. The title of the repo. has been changed
2. Manuel Instalation added.
3. Links for pre-trained models stored on Hobbit Data are provided.
4. How to cite section is added.
5. Regression tests are added.

Page 2 of 2

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.