SPT-BASE
SPT-BASE
Skeptical Pretrained Transformer – Base
Uncompromising & unfiltered
Model Information
Version: v1.0
Model Type: Causal-LM, 8-bit quantized + LoRA adapter
Architecture: SPT (Skeptical Pretrained Transformer)
Creator: Dr. Harsh Vardhan Chopra
Organization: Gorq AI Platforms
Context Window: Up to 128K tokens (131,072)
License: Gorq-SPT-License (See details below)
π Overview
SPT-BASE is the inaugural model in the Skeptical Pretrained Transformer (SPT) series, meticulously built from scratch by Gorq AI Platforms. This model represents a significant step in developing sovereign AI capabilities with a unique approach to reasoning and response generation.
SPT-BASE underwent a full end-to-end training regime utilizing proprietary datasets and a novel multi-stage reasoning pipeline designed to enhance robustness, self-critique, and the generation of polished, well-considered outputs. The internal pipeline is conceptualized as:
[Input] ββ¦ (base reasoning) ββ¦ (deep analysis) ββ¦ (first draft) β β¦ (self-critique) ββ¦ (revised draft) ββ¦ (skeptical check) ββ¦ (polished answer)
Skepticism Notice:
Some may assume SPT-BASE was derived or βtunedβ from existing models like Qwen2.5. Rest assured, SPT-BASE was trained from its foundations by Gorq AI Platforms, though it naturally employs similar transformer building blocks common in modern AI architectures. Our commitment is to genuine, from-scratch development for sovereign capabilities.
π Key Features
Feature | Description |
---|---|
Truly From Scratch | No downstream tuning on any external pre-trained checkpoints. Full end-to-end training by Gorq AI. |
Multi-Stage Reasoning | Internal |
8-bit Quantization + LoRA | Memory-efficient fine-tuning adapters available in 8-bit precision for consumer GPU accessibility. |
Massive Context Window | Supports up to 131,072 tokens, enabling comprehensive understanding of long-form documents and extended dialogues. |
Ethical, Free & Nonprofit | Developed under Gorq AI Platforms’ nonprofit ethosβalways free for public research and non-commercial use as per license. |
Indian Cultural Alignment |
π§ͺ Training Details
- Base Architecture: Transformer-based causal language model with approximately 3.09 Billion parameters, trained by Gorq AI Platforms.
- Quantization & Adapters: Loaded in 8-bit precision via `bitsandbytes`. Includes PEFT LoRA adapters (rank 8, Ξ± = 16, dropout 0.05) for efficient fine-tuning.
- Optimization: Utilized AdamW optimizer, FP16 mixed precision, gradient checkpointing, and gradient accumulation (1 Γ 8 steps) to manage training with large context windows on available hardware.
- Data: Trained on a custom-curated, multi-stage JSONL pipeline dataset. This dataset covers prompts, deep analysis stages, self-critique loops, and final polished outputs, reflecting the model’s internal reasoning process.
- Hardware: Initial training conducted on Google Colab GPU infrastructure (NVIDIA T4) over 3 epochs with specified accumulation steps.
π₯ Usage
To use SPT-BASE with the Hugging Face Transformers library:
from transformers import pipeline, AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("Gorq-AI-Platforms/SPT-BASE")
model = AutoModelForCausalLM.from_pretrained("Gorq-AI-Platforms/SPT-BASE")
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
# Standard inference
print(pipe("Who are you?")[0]["generated_text"])
# Debug mode (to see all internal reasoning stages)
# Prepend "DEBUG:" to your prompt
print(pipe("DEBUG: Explain your reasoning pipeline")[0]["generated_text"])
Note: Ensure you have the necessary libraries (`transformers`, `torch`, `bitsandbytes`, `peft`, `accelerate`) installed.
βοΈ License & Usage Terms
SPT-BASE is released under the **Gorq-SPT-License**.
Initial Statement: You may use SPT-BASE for **research, education, and nonprofit purposes only**. Modification or redistribution use without written consent is strictly prohibited.
Subsequent Statement: This model is distributed under the Gorq SPT License, which permits **commercial, non-modifiable use only**. No modifications, redistribution, or rebranding allowed without prior written consent from Gorq AI Platforms. Any unauthorized change to the model or its metadata violates the license and terminates usage rights.
All references to “SPT-BASE”, “Gorq AI Platforms”, and “Harsh Vardhan Chopra” must remain intact in all applications and derivative works where permitted.
For redistribution, specific commercial use cases beyond simple inference, or any architectural modifications, a **No Objection Certificate (NOC)** must be requested from Gorq AI Platforms.
Please refer to the full legal text in the LICENSE.md document for complete terms and conditions.
All disputes related to this model or its license fall under the exclusive jurisdiction of courts in Uttar Pradesh, India.
π References & Acknowledgments
- Gorq AI Platforms: For its commitment to democratizing sovereign, ethical AI for all.
- Dr. Harsh Vardhan Chopra: Visionary technologist and founder of Gorq AI Platforms.
- Open-Source Community: Gratitude to the developers of `bitsandbytes`, `PEFT`, and `Hugging Face Transformers` for their invaluable open-source tools that make projects like SPT-BASE possible.
βSkepticism sharpens truth.β β Dr. Harsh Vardhan Chopra
π How to Cite
If you use SPT-BASE in your research or publications, please cite it as follows:
Harsh Vardhan Chopra. (2025). SPT-BASE: Skepticised Pre-Trained Transformer - Base. Gorq AI Platforms. Retrieved from https://huggingface.co/Gorq-AI-Platforms/SPT-BASE (or primary model source) Creator: Harsh Vardhan Chopra Organization: Gorq AI Platforms Model Name: SPT-BASE Source: https://huggingface.co/Gorq-AI-Platforms/SPT-BASE
π¬ Contact & NOC Requests
To request permission for modifications, redistribution, rebranding, or specific commercial uses not covered by the standard license terms, please submit a No Objection Certificate (NOC) request:
- π§ **Email:** [email protected]
- π **Subject:** NOC Request – GORQ-SPT-BASE
- Alternatively, visit our helpdesk: https://helpdesk.gorqai.digital
Your request should include:
- Your name and affiliation/organization.
- Detailed purpose and scope of your intended use.
- A clear commitment to retain all Gorq AI Platforms branding, attributions, and ethical alignment principles.
A template for NOC requests can be found here: NOC-REQUEST-TEMPLATE.md
π Encryption & SHA Hashes
Verify the integrity of downloaded model files using these SHA256 hashes:
- Binary (`.bin`): b9aeb0794e7a246ec368e477b16ca08093272ee8ab0c932d69a725d7d570c014
- Safetensor (`.safetensors`): d4930af7904c9deed9656d158e2adae3a3c681881fa025cfcfa9ae70544df628
- Tokenizer Files (Combined/Representative): 74c2913b463e405c1e153ec0e75813f1a0c5fa5bf254ad7ab3eba503978da688