🧠
Model

Sentence Vae

by timbmg gh-model--timbmg--sentence-vae
Nexus Index
47.1 Top 100%
S: Semantic 50
A: Authority 0
P: Popularity 65
R: Recency 98
Q: Quality 50
Tech Context
Vital Performance
0 DL / 30D
0.0%
Audited 47.1 FNI Score
Tiny - Params
- Context
0 Downloads
Model Information Summary
Entity Passport
Registry ID gh-model--timbmg--sentence-vae
Provider github
📜

Cite this model

Academic & Research Attribution

BibTeX
@misc{gh_model__timbmg__sentence_vae,
  author = {timbmg},
  title = {Sentence Vae Model},
  year = {2026},
  howpublished = {\url{https://github.com/timbmg/sentence-vae}},
  note = {Accessed via Free2AITools Knowledge Fortress}
}
APA Style
timbmg. (2026). Sentence Vae [Model]. Free2AITools. https://github.com/timbmg/sentence-vae

đŸ”ŦTechnical Deep Dive

Full Specifications [+]

Quick Commands

🐙 Git Clone
git clone https://github.com/timbmg/sentence-vae

âš–ī¸ Nexus Index V2.0

47.1
TOP 100% SYSTEM IMPACT
Semantic (S) 50
Authority (A) 0
Popularity (P) 65
Recency (R) 98
Quality (Q) 50

đŸ’Ŧ Index Insight

FNI V2.0 for Sentence Vae: Semantic (S:50), Authority (A:0), Popularity (P:65), Recency (R:98), Quality (Q:50).

Free2AITools Nexus Index

Verification Authority

Unbiased Data Node Refresh: VFS Live
---

🚀 What's Next?

Technical Deep Dive

Sentence Variational Autoencoder

PyTorch re-implementation of Generating Sentences from a Continuous Space by Bowman et al. 2015. Model Architecture Note: This implementation does not support LSTM's at the moment, but RNN's and GRU's.

Results

Training

ELBO

ELBO

Negative Log Likelihood

NLL

KL Divergence

KL KL Weight

Performance

Training was stopped after 4 epochs. The true ELBO was optimized for approximately 1 epoch (as can bee see in the graph above). Results are averaged over entire split.

Split NLL KL
Train 99.821 7.944
Validation 103.220 7.346
Test 103.967 7.269

Samples

Sentenes have been obtained after sampling from z ~ N(0, I).

mr . n who was n't n with his own staff and the n n n n n
in the n of the n of the u . s . companies are n't likely to be reached for comment
when they were n in the n and then they were n a n n
but the company said it will be n by the end of the n n and n n
but the company said that it will be n n of the u . s . economy

Interpolating Sentences

Sentenes have been obtained after sampling twice from z ~ N(0, I) and the interpolating the two samples.

the company said it will be n with the exception of the company
but the company said it will be n with the exception of the company ' s shares outstanding
but the company said that the company ' s n n and n n
but the company ' s n n in the past two years ago
but the company ' s n n in the past two years ago
but in the past few years ago that the company ' s n n
but in the past few years ago that they were n't disclosed
but in the past few years ago that they were n't disclosed
but in a statement that they were n't aware of the $ n million in the past few weeks
but in a statement that they were n't paid by the end of the past few weeks

Training

To run the training, please download the Penn Tree Bank data first (download from Tomas Mikolov's webpage). The code expects to find at least ptb.train.txt and ptb.valid.txt in the specified data directory. The data can also be donwloaded with the dowloaddata.sh script.

Then training can be executed with the following command:

text
python3 train.py

The following arguments are available:

--data_dir The path to the directory where PTB data is stored, and auxiliary data files will be stored.
--create_data If provided, new auxiliary data files will be created form the source data.
--max_sequence_length Specifies the cut off of long sentences.
--min_occ If a word occurs less than "min_occ" times in the corpus, it will be replaced by the token.
--test If provided, performance will also be measured on the test set.

-ep, --epochs
-bs, --batch_size
-lr, --learning_rate

-eb, --embedding_size
-rnn, --rnn_type Either 'rnn' or 'gru'.
-hs, --hidden_size
-nl, --num_layers
-bi, --bidirectional
-ls, --latent_size
-wd, --word_dropout Word dropout applied to the input of the Decoder, which means words will be replaced by <unk> with a probability of word_dropout.
-ed, --embedding_dropout Word embedding dropout applied to the input of the Decoder.

-af, --anneal_function Either 'logistic' or 'linear'.
-k, --k Steepness of the logistic annealing function.
-x0, --x0 For 'logistic', this is the mid-point (i.e. when the weight is 0.5); for 'linear' this is the denominator.

-v, --print_every
-tb, --tensorboard_logging If provided, training progress is monitored with tensorboard.
-log, --logdir Directory of log files for tensorboard.
-bin,--save_model_path Directory where to store model checkpoints.

Inference

For obtaining samples and interpolating between senteces, inference.py can be used.

text
python3 inference.py -c $CHECKPOINT -n $NUM_SAMPLES

âš ī¸ Incomplete Data

Some information about this model is not available. Use with Caution - Verify details from the original source before relying on this data.

View Original Source →

📝 Limitations & Considerations

  • â€ĸ Benchmark scores may vary based on evaluation methodology and hardware configuration.
  • â€ĸ VRAM requirements are estimates; actual usage depends on quantization and batch size.
  • â€ĸ FNI scores are relative rankings and may change as new models are added.
  • ⚠ License Unknown: Verify licensing terms before commercial use.

Social Proof

GitHub Repository
592Stars
🔄 Daily sync (03:00 UTC)

AI Summary: Based on GitHub metadata. Not a recommendation.

📊 FNI Methodology 📚 Knowledge Baseâ„šī¸ Verify with original source

đŸ›Ąī¸ Model Transparency Report

Technical metadata sourced from upstream repositories.

Open Metadata

🆔 Identity & Source

id
gh-model--timbmg--sentence-vae
slug
timbmg--sentence-vae
source
github
author
timbmg
license
tags
deep-learning, generative-model, neural-network, nlp, ptb, pytorch, vae, python

âš™ī¸ Technical Specs

architecture
null
params billions
null
context length
null
pipeline tag
other

📊 Engagement & Metrics

downloads
0
stars
592
forks
0

Data indexed from public sources. Updated daily.