| import json |
| import torch |
| from torch.utils.data import IterableDataset, DataLoader |
| from transformers import AutoTokenizer, AutoModelForCausalLM |
| from tqdm import tqdm |
| from math import ceil |
|
|
| class JsonlIterableDataset(IterableDataset): |
| """Sequential streaming dataset for jsonl lines of the form {"text": "..."}.""" |
| def __init__(self, jsonl_path: str, tokenizer, target_token_length: int): |
| super().__init__() |
| self.jsonl_path = jsonl_path |
| self.tokenizer = tokenizer |
| if tokenizer.pad_token is None: |
| tokenizer.pad_token = tokenizer.eos_token |
| self.target_token_length = target_token_length |
|
|
| def __iter__(self): |
| worker_info = torch.utils.data.get_worker_info() |
| if worker_info is None: |
| |
| start, stride = 0, 1 |
| else: |
| |
| start = worker_info.id |
| stride = worker_info.num_workers |
|
|
| with open(self.jsonl_path, "r", encoding="utf-8") as f: |
| for idx, line in enumerate(f): |
| if idx % stride != start: |
| continue |
| data = json.loads(line) |
| text = data["text"] |
|
|
| tokens = self.tokenizer( |
| text, |
| truncation=True, |
| padding="max_length", |
| max_length=self.target_token_length, |
| return_tensors="pt" |
| ) |
| yield { |
| "input_ids": tokens["input_ids"].squeeze(0), |
| "attention_mask": tokens["attention_mask"].squeeze(0), |
| } |
|
|
|
|
| def calculate_negative_log_likelihood( |
| model_path: str, |
| jsonl_path: str, |
| target_token_length: int, |
| batch_size: int = 8, |
| device: str = "cuda" if torch.cuda.is_available() else "cpu", |
| num_workers: int = 2, |
| num_samples: int = None, |
| ) -> torch.Tensor: |
| """ |
| Streaming, batched NLL computation for a large jsonl dataset using deterministic sequential access. |
| """ |
| tokenizer = AutoTokenizer.from_pretrained(model_path, device_map="auto", trust_remote_code=True) |
| model = AutoModelForCausalLM.from_pretrained(model_path, device_map="auto", torch_dtype="auto", |
| attn_implementation="flash_attention_2" if "TinyLlama" not in model_path else "sdpa", trust_remote_code=True) |
| model.eval() |
|
|
| dataset = JsonlIterableDataset(jsonl_path, tokenizer, target_token_length) |
| dataloader = DataLoader(dataset, batch_size=batch_size, num_workers=num_workers) |
|
|
| entropies = [] |
|
|
| for i, batch in enumerate(tqdm(dataloader, total=ceil(num_samples / batch_size) if num_samples is not None else None, |
| desc=f"Calculating Entropy for {model_path.split('/')[-1]}")): |
| if i % 100 == 0: torch.cuda.empty_cache() |
|
|
| input_ids = batch["input_ids"].to(device) |
| attention_mask = batch["attention_mask"].to(device) |
|
|
| with torch.no_grad(): |
| outputs = model(input_ids=input_ids, attention_mask=attention_mask, use_cache=False) |
| logits = outputs.logits |
|
|
| |
| logits = torch.softmax(logits[:, :, :len(tokenizer)].to(dtype=torch.float32), dim=-1) |
| effective_probs = torch.gather(logits[:, :target_token_length, :], -1, input_ids[:, 1:].unsqueeze(-1)).squeeze(-1) |
| entropy = -torch.log2(effective_probs) |
| entropy[attention_mask[:, 1:] == 0] = torch.nan |
|
|
| entropies.append(entropy.cpu()) |
|
|
| return torch.cat(entropies, dim=0) |