WFGY/wfgy_core.py
2025-06-11 17:49:33 +08:00

38 lines
1.5 KiB
Python

import torch
from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM
class WFGYRunner:
def __init__(self, model_id="sshleifer/tiny-gpt2", use_remote=False):
self.use_remote = use_remote
self.device = "cuda" if torch.cuda.is_available() else "cpu"
self.model_id = model_id
if self.use_remote:
raise NotImplementedError("Remote mode is currently disabled in this test version.")
else:
self.tokenizer = AutoTokenizer.from_pretrained(model_id)
self.model = AutoModelForCausalLM.from_pretrained(model_id)
self.pipe = pipeline(
"text-generation",
model=self.model,
tokenizer=self.tokenizer,
device=0 if self.device == "cuda" else -1
)
def run(self, prompt, max_new_tokens=100, temperature=0.7):
print("╭──────────────────────────────╮")
print("│ 🤖 INITIATING WFGY CORE │")
print("│ ⚙️ MODULE: Semantic Boost │")
print("╰──────────────────────────────╯\n")
print("=== Prompt ===")
print(prompt)
result = self.pipe(
prompt,
max_new_tokens=max_new_tokens,
temperature=temperature
)[0]["generated_text"]
print("\n=== Output ===")
print(result.strip())
return result.strip()