| import re |
| import torch |
| from difflib import SequenceMatcher |
| from transformers import T5Tokenizer, T5ForConditionalGeneration |
| from transformers.utils import logging as hf_logging |
|
|
| hf_logging.set_verbosity_error() |
|
|
| MODEL_DIR = "t5-viet-qg-finetuned" |
|
|
| PATTERN = re.compile( |
| r""" |
| “([^”]{3,120})” |
| | "([^"]{3,120})" |
| | \b(?:là|gồm|do|theo)\s+([^,.;:\n]{3,120}) |
| | \b\d{4}\b |
| | \b(?:Điều|Khoản)\s+\d+\b |
| """, |
| re.VERBOSE | re.IGNORECASE, |
| ) |
|
|
|
|
| def norm(s: str) -> str: |
| return re.sub(r"\s+", " ", s).strip() |
|
|
|
|
| def is_dup(q: str, qs: list[str], thr: float = 0.85) -> bool: |
| ql = q.lower() |
| for x in qs: |
| if SequenceMatcher(None, ql, x.lower()).ratio() >= thr: |
| return True |
| return False |
|
|
|
|
| def extract_answers(ctx: str, max_n: int = 60) -> list[str]: |
| ctx = norm(ctx) |
| answers, seen = [], set() |
|
|
| for m in PATTERN.finditer(ctx): |
| for g in m.groups(): |
| if not g: |
| continue |
| g = norm(g) |
| k = g.lower() |
| if 3 <= len(g) <= 120 and k not in seen: |
| seen.add(k) |
| answers.append(g) |
| if len(answers) >= max_n: |
| return answers |
|
|
| if len(answers) < 8: |
| for i in range(0, min(len(ctx), 500), 60): |
| ch = norm(ctx[i : i + 60]) |
| k = ch.lower() |
| if len(ch) >= 15 and k not in seen: |
| seen.add(k) |
| answers.append(ch) |
| if len(answers) >= max_n: |
| break |
|
|
| if not answers and ctx: |
| answers = [ctx[:120]] |
|
|
| return answers |
|
|
|
|
| def load_model(): |
| tok = T5Tokenizer.from_pretrained(MODEL_DIR) |
| mdl = T5ForConditionalGeneration.from_pretrained(MODEL_DIR) |
|
|
| dev = "cuda" if torch.cuda.is_available() else "cpu" |
| try: |
| mdl = mdl.to(dev) |
| except RuntimeError: |
| dev = "cpu" |
| try: |
| torch.cuda.empty_cache() |
| except Exception: |
| pass |
| mdl = mdl.to(dev) |
|
|
| mdl.eval() |
| return tok, mdl, dev |
|
|
|
|
| tokenizer, model, device = load_model() |
|
|
|
|
| def generate_questions(context: str, n: int = 20) -> list[str]: |
| ctx = norm(context) |
| answers = extract_answers(ctx, max_n=80) |
| questions = [] |
|
|
| gen_cfg = dict( |
| do_sample=True, |
| top_k=80, |
| top_p=0.98, |
| temperature=1.05, |
| max_new_tokens=72, |
| no_repeat_ngram_size=3, |
| repetition_penalty=1.08, |
| ) |
|
|
| num_ret = 8 if n <= 20 else 10 |
|
|
| def run_prompt(ans: str, rounds: int): |
| nonlocal gen_cfg |
| prompt = f"answer: {ans}\ncontext: {ctx}\nquestion:" |
| inputs = tokenizer(prompt, return_tensors="pt", truncation=True, max_length=512).to(device) |
|
|
| for _ in range(rounds): |
| outs = model.generate(**inputs, num_return_sequences=num_ret, **gen_cfg) |
| added = 0 |
| for o in outs: |
| q = norm(tokenizer.decode(o, skip_special_tokens=True)) |
| if not q: |
| continue |
| if not q.endswith("?"): |
| q += "?" |
| if len(q) >= 6 and not is_dup(q, questions, thr=0.85): |
| questions.append(q) |
| added += 1 |
| if len(questions) >= n: |
| return |
| if added == 0: |
| gen_cfg["temperature"] = min(1.25, gen_cfg["temperature"] + 0.05) |
| gen_cfg["top_p"] = min(0.995, gen_cfg["top_p"] + 0.005) |
|
|
| with torch.inference_mode(): |
| for ans in answers: |
| if len(questions) >= n: |
| break |
| run_prompt(ans, rounds=6) |
|
|
| if len(questions) < n: |
| run_prompt(ctx[:120], rounds=12) |
|
|
| return questions[:n] |
|
|
|
|
| if __name__ == "__main__": |
| ctx = input("\nNhập đoạn văn bản:\n").strip() |
| try: |
| n = int((input("\nNhập số lượng câu hỏi cần sinh: ").strip() or "")) |
| except ValueError: |
| n = 20 |
|
|
| n = max(1, min(n, 200)) |
| qs = generate_questions(ctx, n) |
|
|
| print("\nCác câu hỏi sinh ra:") |
| for i, q in enumerate(qs, 1): |
| print(f"{i}. {q}") |
|
|