🧠
Model

Rally 4b

by thomasjvu hf-model--thomasjvu--rally-4b
Nexus Index
35.6 Top 100%
S: Semantic 50
A: Authority 0
P: Popularity 0
R: Recency 97
Q: Quality 50
Tech Context
4 Params
4.096K Ctx
Vital Performance
0 DL / 30D
0.0%
Audited 35.6 FNI Score
4B Params
4k Context
0 Downloads
8G GPU ~5GB Est. VRAM
Model Information Summary
Entity Passport
Registry ID hf-model--thomasjvu--rally-4b
Provider huggingface
💾

Compute Threshold

~4.3GB VRAM

Interactive
Analyze Hardware
â–ŧ

* Static estimation for 4-Bit Quantization.

📜

Cite this model

Academic & Research Attribution

BibTeX
@misc{hf_model__thomasjvu__rally_4b,
  author = {thomasjvu},
  title = {Rally 4b Model},
  year = {2026},
  howpublished = {\url{https://huggingface.co/thomasjvu/rally-4b}},
  note = {Accessed via Free2AITools Knowledge Fortress}
}
APA Style
thomasjvu. (2026). Rally 4b [Model]. Free2AITools. https://huggingface.co/thomasjvu/rally-4b

đŸ”ŦTechnical Deep Dive

Full Specifications [+]

Quick Commands

đŸĻ™ Ollama Run
ollama run rally-4b
🤗 HF Download
huggingface-cli download thomasjvu/rally-4b
đŸ“Ļ Install Lib
pip install -U transformers

âš–ī¸ Nexus Index V2.0

35.6
TOP 100% SYSTEM IMPACT
Semantic (S) 50
Authority (A) 0
Popularity (P) 0
Recency (R) 97
Quality (Q) 50

đŸ’Ŧ Index Insight

FNI V2.0 for Rally 4b: Semantic (S:50), Authority (A:0), Popularity (P:0), Recency (R:97), Quality (Q:50).

Free2AITools Nexus Index

Verification Authority

Unbiased Data Node Refresh: VFS Live
---

🚀 What's Next?

Technical Deep Dive

rally-4b

Browser-oriented ONNX export of a Gemma 4 Heretic checkpoint packaged for WebGPU / Transformers.js.

Capabilities

  • Supported inputs: text, image
  • Not included in this package: audio, video

Version Notes

  • This is the stable v1 browser package.
  • For the enhanced multimodal variant, see thomasjvu/rally-4b-v2.
  • The v2 package adds audio, video for this model family.

Provenance

  • Source model: coder3101/gemma-4-E4B-it-heretic
  • Base model for inherited processor assets: google/gemma-4-E4B-it
  • Architecture family: gemma4_conditional_generation
  • Expected architecture: Gemma4ForConditionalGeneration
  • Target dtype: q4f16
  • Target device: webgpu

Expected ONNX Sessions

  • vision_encoder_q4f16.onnx
  • embed_tokens_q4f16.onnx
  • decoder_model_merged_q4f16.onnx

Usage

Load this repo with Transformers.js using the model's transformers.js_config metadata and WebGPU backend.

âš ī¸ Incomplete Data

Some information about this model is not available. Use with Caution - Verify details from the original source before relying on this data.

View Original Source →

📝 Limitations & Considerations

  • â€ĸ Benchmark scores may vary based on evaluation methodology and hardware configuration.
  • â€ĸ VRAM requirements are estimates; actual usage depends on quantization and batch size.
  • â€ĸ FNI scores are relative rankings and may change as new models are added.
  • ⚠ License Unknown: Verify licensing terms before commercial use.
🔄 Daily sync (03:00 UTC)

AI Summary: Based on Hugging Face metadata. Not a recommendation.

📊 FNI Methodology 📚 Knowledge Baseâ„šī¸ Verify with original source

đŸ›Ąī¸ Model Transparency Report

Technical metadata sourced from upstream repositories.

Open Metadata

🆔 Identity & Source

id
hf-model--thomasjvu--rally-4b
slug
thomasjvu--rally-4b
source
huggingface
author
thomasjvu
license
tags
transformers, onnx, gemma4, image-text-to-text, webgpu, transformers.js, quantized, conversational, base_model:coder3101/gemma-4-e4b-it-heretic, endpoints_compatible, region:us

âš™ī¸ Technical Specs

architecture
null
params billions
4
context length
4,096
pipeline tag
image-text-to-text
vram gb
4.3
vram is estimated
true
vram formula
VRAM ≈ (params * 0.75) + 0.8GB (KV) + 0.5GB (OS)

📊 Engagement & Metrics

downloads
0
stars
0
forks
0

Data indexed from public sources. Updated daily.