MCGPT-1: Mixture of Experts (MoE) Language Model

MCGPT-1 is a custom-built MoE model developed by TopAI-IL. It is designed to demonstrate specialized knowledge in Minecraft, Reddit-style conversations, and model self-identity.

Model Details

  • Architecture: Mixture of Experts (MoE)
  • Total Experts: 4
  • Layers: 4
  • Attention Heads: 8
  • Hidden Size: 256
  • Training Domains: 1. Identity (TopAI-IL)
    1. Minecraft Technical Data & Guides
    2. Reddit/Web Slang & Conversations
    3. General Hebrew/English Knowledge
    4. Instructions Syntetic-Data

How to use

This model uses a custom architecture (mcgpt). To run inference, ensure you include the architecture class in your code or use the trust_remote_code=True flag if the modeling script is provided.

Use example

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model_id = "TopAI-1/MCGPT-1" 

# 2. load the model and tokienizer
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
  model_id, 
  trust_remote_code=True, 
  torch_dtype=torch.float32
)

# 3. GPU If have
device = "cuda" if torch.cuda.is_available() else "cpu"
model.to(device)
model.eval()

# 4. fast text generation
def generate(prompt, max_new_tokens=50):
  inputs = tokenizer(prompt, return_tensors="pt").to(device)
  with torch.no_grad():
      outputs = model.generate(
          **inputs, 
          max_new_tokens=max_new_tokens, 
          do_sample=True, 
          top_k=50, 
          temperature=0.8,
          pad_token_id=tokenizer.eos_token_id
      )
  return tokenizer.decode(outputs[0], skip_special_tokens=True)

# Infrence Check
print("Testing MCGPT-1 from Hub:")
prompt = "use the following search parameters to narrow your results: e.g."
print(generate(prompt))

Capabilities

The model successfully identifies itself as MCGPT-1 and can switch between experts based on the prompt (e.g., providing Minecraft-related advice when prompted with "help").

Developed by Raziel @ TopAI-IL (2026)

Downloads last month
307
Safetensors
Model size
22.4M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for TopAI-1/MCGPT-1

Finetunes
1 model

Datasets used to train TopAI-1/MCGPT-1