Add qwen files.
This commit is contained in:
parent
3a4e99f7e3
commit
9b90c607e0
|
@ -0,0 +1,37 @@
|
||||||
|
{
|
||||||
|
"architectures": [
|
||||||
|
"QWenLMHeadModel"
|
||||||
|
],
|
||||||
|
"auto_map": {
|
||||||
|
"AutoConfig": "configuration_qwen.QWenConfig",
|
||||||
|
"AutoModelForCausalLM": "modeling_qwen.QWenLMHeadModel"
|
||||||
|
},
|
||||||
|
"attn_dropout_prob": 0.0,
|
||||||
|
"bf16": false,
|
||||||
|
"emb_dropout_prob": 0.0,
|
||||||
|
"fp16": false,
|
||||||
|
"fp32": false,
|
||||||
|
"hidden_size": 2048,
|
||||||
|
"intermediate_size": 11008,
|
||||||
|
"initializer_range": 0.02,
|
||||||
|
"kv_channels": 128,
|
||||||
|
"layer_norm_epsilon": 1e-06,
|
||||||
|
"max_position_embeddings": 8192,
|
||||||
|
"model_type": "qwen",
|
||||||
|
"no_bias": true,
|
||||||
|
"num_attention_heads": 16,
|
||||||
|
"num_hidden_layers": 24,
|
||||||
|
"onnx_safe": null,
|
||||||
|
"rotary_emb_base": 10000,
|
||||||
|
"rotary_pct": 1.0,
|
||||||
|
"scale_attn_weights": true,
|
||||||
|
"seq_length": 8192,
|
||||||
|
"tie_word_embeddings": false,
|
||||||
|
"tokenizer_class": "QWenTokenizer",
|
||||||
|
"transformers_version": "4.32.0",
|
||||||
|
"use_cache": true,
|
||||||
|
"use_dynamic_ntk": true,
|
||||||
|
"use_flash_attn": "auto",
|
||||||
|
"use_logn_attn": true,
|
||||||
|
"vocab_size": 151936
|
||||||
|
}
|
|
@ -0,0 +1,5 @@
|
||||||
|
{
|
||||||
|
"framework": "pytorch",
|
||||||
|
"task": "chat",
|
||||||
|
"allow_remote": true
|
||||||
|
}
|
|
@ -0,0 +1,71 @@
|
||||||
|
# Copyright (c) Alibaba Cloud.
|
||||||
|
#
|
||||||
|
# This source code is licensed under the license found in the
|
||||||
|
# LICENSE file in the root directory of this source tree.
|
||||||
|
|
||||||
|
from transformers import PretrainedConfig
|
||||||
|
|
||||||
|
|
||||||
|
class QWenConfig(PretrainedConfig):
|
||||||
|
model_type = "qwen"
|
||||||
|
keys_to_ignore_at_inference = ["past_key_values"]
|
||||||
|
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
vocab_size=151936,
|
||||||
|
hidden_size=4096,
|
||||||
|
num_hidden_layers=32,
|
||||||
|
num_attention_heads=32,
|
||||||
|
emb_dropout_prob=0.0,
|
||||||
|
attn_dropout_prob=0.0,
|
||||||
|
layer_norm_epsilon=1e-6,
|
||||||
|
initializer_range=0.02,
|
||||||
|
max_position_embeddings=8192,
|
||||||
|
scale_attn_weights=True,
|
||||||
|
use_cache=True,
|
||||||
|
bf16=False,
|
||||||
|
fp16=False,
|
||||||
|
fp32=False,
|
||||||
|
kv_channels=128,
|
||||||
|
rotary_pct=1.0,
|
||||||
|
rotary_emb_base=10000,
|
||||||
|
use_dynamic_ntk=True,
|
||||||
|
use_logn_attn=True,
|
||||||
|
use_flash_attn="auto",
|
||||||
|
intermediate_size=22016,
|
||||||
|
no_bias=True,
|
||||||
|
tie_word_embeddings=False,
|
||||||
|
use_cache_quantization=False,
|
||||||
|
use_cache_kernel=False,
|
||||||
|
softmax_in_fp32=False,
|
||||||
|
**kwargs,
|
||||||
|
):
|
||||||
|
self.vocab_size = vocab_size
|
||||||
|
self.hidden_size = hidden_size
|
||||||
|
self.intermediate_size = intermediate_size
|
||||||
|
self.num_hidden_layers = num_hidden_layers
|
||||||
|
self.num_attention_heads = num_attention_heads
|
||||||
|
self.emb_dropout_prob = emb_dropout_prob
|
||||||
|
self.attn_dropout_prob = attn_dropout_prob
|
||||||
|
self.layer_norm_epsilon = layer_norm_epsilon
|
||||||
|
self.initializer_range = initializer_range
|
||||||
|
self.scale_attn_weights = scale_attn_weights
|
||||||
|
self.use_cache = use_cache
|
||||||
|
self.max_position_embeddings = max_position_embeddings
|
||||||
|
self.bf16 = bf16
|
||||||
|
self.fp16 = fp16
|
||||||
|
self.fp32 = fp32
|
||||||
|
self.kv_channels = kv_channels
|
||||||
|
self.rotary_pct = rotary_pct
|
||||||
|
self.rotary_emb_base = rotary_emb_base
|
||||||
|
self.use_dynamic_ntk = use_dynamic_ntk
|
||||||
|
self.use_logn_attn = use_logn_attn
|
||||||
|
self.use_flash_attn = use_flash_attn
|
||||||
|
self.no_bias = no_bias
|
||||||
|
self.use_cache_quantization = use_cache_quantization
|
||||||
|
self.use_cache_kernel = use_cache_kernel
|
||||||
|
self.softmax_in_fp32 = softmax_in_fp32
|
||||||
|
super().__init__(
|
||||||
|
tie_word_embeddings=tie_word_embeddings,
|
||||||
|
**kwargs
|
||||||
|
)
|
|
@ -0,0 +1,55 @@
|
||||||
|
from torch.utils import cpp_extension
|
||||||
|
import pathlib
|
||||||
|
import os
|
||||||
|
import subprocess
|
||||||
|
|
||||||
|
def _get_cuda_bare_metal_version(cuda_dir):
|
||||||
|
raw_output = subprocess.check_output([cuda_dir + "/bin/nvcc", "-V"],
|
||||||
|
universal_newlines=True)
|
||||||
|
output = raw_output.split()
|
||||||
|
release_idx = output.index("release") + 1
|
||||||
|
release = output[release_idx].split(".")
|
||||||
|
bare_metal_major = release[0]
|
||||||
|
bare_metal_minor = release[1][0]
|
||||||
|
|
||||||
|
return raw_output, bare_metal_major, bare_metal_minor
|
||||||
|
|
||||||
|
def _create_build_dir(buildpath):
|
||||||
|
try:
|
||||||
|
os.mkdir(buildpath)
|
||||||
|
except OSError:
|
||||||
|
if not os.path.isdir(buildpath):
|
||||||
|
print(f"Creation of the build directory {buildpath} failed")
|
||||||
|
|
||||||
|
# Check if cuda 11 is installed for compute capability 8.0
|
||||||
|
cc_flag = []
|
||||||
|
_, bare_metal_major, bare_metal_minor = _get_cuda_bare_metal_version(cpp_extension.CUDA_HOME)
|
||||||
|
if int(bare_metal_major) >= 11:
|
||||||
|
cc_flag.append('-gencode')
|
||||||
|
cc_flag.append('arch=compute_80,code=sm_80')
|
||||||
|
if int(bare_metal_minor) >= 7:
|
||||||
|
cc_flag.append('-gencode')
|
||||||
|
cc_flag.append('arch=compute_90,code=sm_90')
|
||||||
|
|
||||||
|
# Build path
|
||||||
|
srcpath = pathlib.Path(__file__).parent.absolute()
|
||||||
|
buildpath = srcpath / 'build'
|
||||||
|
_create_build_dir(buildpath)
|
||||||
|
|
||||||
|
def _cpp_extention_load_helper(name, sources, extra_cuda_flags):
|
||||||
|
return cpp_extension.load(
|
||||||
|
name=name,
|
||||||
|
sources=sources,
|
||||||
|
build_directory=buildpath,
|
||||||
|
extra_cflags=['-O3', ],
|
||||||
|
extra_cuda_cflags=['-O3',
|
||||||
|
'-gencode', 'arch=compute_70,code=sm_70',
|
||||||
|
'--use_fast_math'] + extra_cuda_flags + cc_flag,
|
||||||
|
verbose=1
|
||||||
|
)
|
||||||
|
|
||||||
|
extra_flags = []
|
||||||
|
|
||||||
|
cache_autogptq_cuda_256_sources = ["./cache_autogptq_cuda_256.cpp",
|
||||||
|
"./cache_autogptq_cuda_kernel_256.cu"]
|
||||||
|
cache_autogptq_cuda_256 = _cpp_extention_load_helper("cache_autogptq_cuda_256", cache_autogptq_cuda_256_sources, extra_flags)
|
21
qwen/demo.py
21
qwen/demo.py
|
@ -1,11 +1,30 @@
|
||||||
|
|
||||||
|
import torch
|
||||||
from modelscope import snapshot_download
|
from modelscope import snapshot_download
|
||||||
from transformers import AutoModelForCausalLM, AutoTokenizer
|
from transformers import AutoModelForCausalLM, AutoTokenizer
|
||||||
from transformers.generation import GenerationConfig
|
from transformers.generation import GenerationConfig
|
||||||
|
from transformers import AutoConfig
|
||||||
|
|
||||||
|
from modeling_qwen import QWenLMHeadModel
|
||||||
|
|
||||||
|
seed = 4321
|
||||||
|
torch.manual_seed(seed)
|
||||||
|
torch.cuda.manual_seed_all(seed)
|
||||||
|
|
||||||
model_dir = snapshot_download("qwen/Qwen-1_8B-Chat")
|
model_dir = snapshot_download("qwen/Qwen-1_8B-Chat")
|
||||||
|
|
||||||
|
config, kwargs = AutoConfig.from_pretrained(
|
||||||
|
model_dir,
|
||||||
|
return_unused_kwargs=True,
|
||||||
|
trust_remote_code=True,
|
||||||
|
code_revision=None,
|
||||||
|
_commit_hash=None,
|
||||||
|
)
|
||||||
|
model = QWenLMHeadModel(config)
|
||||||
|
|
||||||
|
|
||||||
tokenizer = AutoTokenizer.from_pretrained(model_dir, trust_remote_code=True)
|
tokenizer = AutoTokenizer.from_pretrained(model_dir, trust_remote_code=True)
|
||||||
model = AutoModelForCausalLM.from_pretrained(
|
model = model.from_pretrained(
|
||||||
model_dir, device_map="auto", trust_remote_code=True
|
model_dir, device_map="auto", trust_remote_code=True
|
||||||
).eval()
|
).eval()
|
||||||
|
|
||||||
|
|
|
@ -0,0 +1,12 @@
|
||||||
|
{
|
||||||
|
"chat_format": "chatml",
|
||||||
|
"eos_token_id": 151643,
|
||||||
|
"pad_token_id": 151643,
|
||||||
|
"max_window_size": 6144,
|
||||||
|
"max_new_tokens": 512,
|
||||||
|
"do_sample": true,
|
||||||
|
"top_k": 0,
|
||||||
|
"top_p": 0.8,
|
||||||
|
"repetition_penalty": 1.1,
|
||||||
|
"transformers_version": "4.31.0"
|
||||||
|
}
|
|
@ -0,0 +1,202 @@
|
||||||
|
{
|
||||||
|
"metadata": {
|
||||||
|
"total_size": 3673657344
|
||||||
|
},
|
||||||
|
"weight_map": {
|
||||||
|
"lm_head.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.0.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.0.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.0.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.0.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.0.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.0.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.0.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.0.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.1.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.1.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.1.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.1.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.1.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.1.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.1.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.1.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.10.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.10.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.10.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.10.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.10.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.10.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.10.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.10.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.11.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.11.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.11.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.11.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.11.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.11.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.11.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.11.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.12.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.12.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.12.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.12.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.12.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.12.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.12.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.12.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.13.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.13.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.13.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.13.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.13.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.13.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.13.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.13.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.14.attn.c_attn.bias": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.14.attn.c_attn.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.14.attn.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.14.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.14.ln_2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.14.mlp.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.14.mlp.w1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.14.mlp.w2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.15.attn.c_attn.bias": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.15.attn.c_attn.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.15.attn.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.15.ln_1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.15.ln_2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.15.mlp.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.15.mlp.w1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.15.mlp.w2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.16.attn.c_attn.bias": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.16.attn.c_attn.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.16.attn.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.16.ln_1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.16.ln_2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.16.mlp.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.16.mlp.w1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.16.mlp.w2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.17.attn.c_attn.bias": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.17.attn.c_attn.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.17.attn.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.17.ln_1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.17.ln_2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.17.mlp.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.17.mlp.w1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.17.mlp.w2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.18.attn.c_attn.bias": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.18.attn.c_attn.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.18.attn.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.18.ln_1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.18.ln_2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.18.mlp.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.18.mlp.w1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.18.mlp.w2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.19.attn.c_attn.bias": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.19.attn.c_attn.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.19.attn.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.19.ln_1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.19.ln_2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.19.mlp.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.19.mlp.w1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.19.mlp.w2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.2.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.2.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.2.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.2.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.2.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.2.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.2.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.2.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.20.attn.c_attn.bias": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.20.attn.c_attn.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.20.attn.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.20.ln_1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.20.ln_2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.20.mlp.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.20.mlp.w1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.20.mlp.w2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.21.attn.c_attn.bias": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.21.attn.c_attn.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.21.attn.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.21.ln_1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.21.ln_2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.21.mlp.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.21.mlp.w1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.21.mlp.w2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.22.attn.c_attn.bias": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.22.attn.c_attn.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.22.attn.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.22.ln_1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.22.ln_2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.22.mlp.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.22.mlp.w1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.22.mlp.w2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.23.attn.c_attn.bias": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.23.attn.c_attn.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.23.attn.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.23.ln_1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.23.ln_2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.23.mlp.c_proj.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.23.mlp.w1.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.23.mlp.w2.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.h.3.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.3.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.3.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.3.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.3.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.3.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.3.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.3.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.4.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.4.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.4.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.4.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.4.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.4.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.4.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.4.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.5.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.5.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.5.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.5.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.5.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.5.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.5.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.5.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.6.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.6.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.6.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.6.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.6.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.6.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.6.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.6.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.7.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.7.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.7.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.7.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.7.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.7.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.7.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.7.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.8.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.8.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.8.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.8.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.8.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.8.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.8.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.8.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.9.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.9.attn.c_attn.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.9.attn.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.9.ln_1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.9.ln_2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.9.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.9.mlp.w1.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.h.9.mlp.w2.weight": "model-00001-of-00002.safetensors",
|
||||||
|
"transformer.ln_f.weight": "model-00002-of-00002.safetensors",
|
||||||
|
"transformer.wte.weight": "model-00001-of-00002.safetensors"
|
||||||
|
}
|
||||||
|
}
|
|
@ -40,8 +40,8 @@ SUPPORT_FP16 = SUPPORT_CUDA and torch.cuda.get_device_capability(0)[0] >= 7
|
||||||
SUPPORT_TORCH2 = hasattr(torch, '__version__') and int(torch.__version__.split(".")[0]) >= 2
|
SUPPORT_TORCH2 = hasattr(torch, '__version__') and int(torch.__version__.split(".")[0]) >= 2
|
||||||
|
|
||||||
|
|
||||||
from .configuration_qwen import QWenConfig
|
from configuration_qwen import QWenConfig
|
||||||
from .qwen_generation_utils import (
|
from qwen_generation_utils import (
|
||||||
HistoryType,
|
HistoryType,
|
||||||
make_context,
|
make_context,
|
||||||
decode_tokens,
|
decode_tokens,
|
||||||
|
@ -520,7 +520,9 @@ class QWenAttention(nn.Module):
|
||||||
|
|
||||||
if not self.use_cache_quantization and SUPPORT_TORCH2:
|
if not self.use_cache_quantization and SUPPORT_TORCH2:
|
||||||
if attention_mask is not None:
|
if attention_mask is not None:
|
||||||
attention_mask = attention_mask.expand(-1, -1, query.size(2), -1)
|
attention_mask = attention_mask.expand(
|
||||||
|
-1, -1, causal_mask.size(2), -1
|
||||||
|
)
|
||||||
if causal_mask is not None:
|
if causal_mask is not None:
|
||||||
attention_mask = attention_mask.masked_fill(~causal_mask, torch.finfo(query.dtype).min)
|
attention_mask = attention_mask.masked_fill(~causal_mask, torch.finfo(query.dtype).min)
|
||||||
else:
|
else:
|
||||||
|
@ -1328,14 +1330,14 @@ def apply_rotary_pos_emb(t, freqs):
|
||||||
t (tensor(batch_size, seq_len, n_head, head_dim)):
|
t (tensor(batch_size, seq_len, n_head, head_dim)):
|
||||||
the input embedding/hidden states
|
the input embedding/hidden states
|
||||||
freqs (list[tensor(1, seq_len, 1, rotary_dim), tensor(1, seq_len, 1, rotary_dim)]):
|
freqs (list[tensor(1, seq_len, 1, rotary_dim), tensor(1, seq_len, 1, rotary_dim)]):
|
||||||
the cached cos/sin position embeddings
|
the cached cos/sin position embeddings
|
||||||
"""
|
"""
|
||||||
rot_dim = freqs[0].shape[-1]
|
rot_dim = freqs[0].shape[-1]
|
||||||
cos, sin = freqs
|
cos, sin = freqs
|
||||||
t_float = t.float()
|
t_float = t.float()
|
||||||
if apply_rotary_emb_func is not None and t.is_cuda:
|
if apply_rotary_emb_func is not None and t.is_cuda:
|
||||||
# apply_rotary_emb in flash_attn requires cos/sin to be of
|
# apply_rotary_emb in flash_attn requires cos/sin to be of
|
||||||
# shape (seqlen, rotary_dim / 2) and apply rotary embedding
|
# shape (seqlen, rotary_dim / 2) and apply rotary embedding
|
||||||
# to the first rotary_dim of the input
|
# to the first rotary_dim of the input
|
||||||
cos = cos.squeeze(0).squeeze(1)[:, : rot_dim // 2]
|
cos = cos.squeeze(0).squeeze(1)[:, : rot_dim // 2]
|
||||||
sin = sin.squeeze(0).squeeze(1)[:, : rot_dim // 2]
|
sin = sin.squeeze(0).squeeze(1)[:, : rot_dim // 2]
|
||||||
|
@ -1360,4 +1362,4 @@ class RMSNorm(torch.nn.Module):
|
||||||
return rms_norm(x, self.weight, self.eps)
|
return rms_norm(x, self.weight, self.eps)
|
||||||
else:
|
else:
|
||||||
output = self._norm(x.float()).type_as(x)
|
output = self._norm(x.float()).type_as(x)
|
||||||
return output * self.weight
|
return output * self.weight
|
||||||
|
|
|
@ -0,0 +1,416 @@
|
||||||
|
# Copyright (c) Alibaba Cloud.
|
||||||
|
#
|
||||||
|
# This source code is licensed under the license found in the
|
||||||
|
# LICENSE file in the root directory of this source tree.
|
||||||
|
|
||||||
|
"""Generation support."""
|
||||||
|
|
||||||
|
from typing import Tuple, List, Union, Iterable
|
||||||
|
|
||||||
|
import numpy as np
|
||||||
|
import torch
|
||||||
|
import torch.nn.functional as F
|
||||||
|
from transformers import PreTrainedTokenizer
|
||||||
|
from transformers import logging
|
||||||
|
from transformers.generation import LogitsProcessor
|
||||||
|
|
||||||
|
logger = logging.get_logger(__name__)
|
||||||
|
|
||||||
|
# Types.
|
||||||
|
HistoryType = List[Tuple[str, str]]
|
||||||
|
TokensType = List[int]
|
||||||
|
BatchTokensType = List[List[int]]
|
||||||
|
|
||||||
|
|
||||||
|
def pad_batch(batch: BatchTokensType, pad_id: int, seq_length: int) -> BatchTokensType:
|
||||||
|
for tokens in batch:
|
||||||
|
context_length = len(tokens)
|
||||||
|
if context_length < seq_length:
|
||||||
|
tokens.extend([pad_id] * (seq_length - context_length))
|
||||||
|
return batch
|
||||||
|
|
||||||
|
|
||||||
|
def get_ltor_masks_and_position_ids(
|
||||||
|
data,
|
||||||
|
eod_token,
|
||||||
|
reset_position_ids,
|
||||||
|
reset_attention_mask,
|
||||||
|
eod_mask_loss,
|
||||||
|
):
|
||||||
|
"""Build masks and position id for left to right model."""
|
||||||
|
|
||||||
|
# Extract batch size and sequence length.
|
||||||
|
micro_batch_size, seq_length = data.size()
|
||||||
|
|
||||||
|
# Attention mask (lower triangular).
|
||||||
|
if reset_attention_mask:
|
||||||
|
att_mask_batch = micro_batch_size
|
||||||
|
else:
|
||||||
|
att_mask_batch = 1
|
||||||
|
attention_mask = torch.tril(
|
||||||
|
torch.ones((att_mask_batch, seq_length, seq_length), device=data.device)
|
||||||
|
).view(att_mask_batch, 1, seq_length, seq_length)
|
||||||
|
|
||||||
|
# Loss mask.
|
||||||
|
loss_mask = torch.ones(data.size(), dtype=torch.float, device=data.device)
|
||||||
|
if eod_mask_loss:
|
||||||
|
loss_mask[data == eod_token] = 0.0
|
||||||
|
|
||||||
|
# Position ids.
|
||||||
|
position_ids = torch.arange(seq_length, dtype=torch.long, device=data.device)
|
||||||
|
position_ids = position_ids.unsqueeze(0).expand_as(data)
|
||||||
|
# We need to clone as the ids will be modifed based on batch index.
|
||||||
|
if reset_position_ids:
|
||||||
|
position_ids = position_ids.clone()
|
||||||
|
|
||||||
|
if reset_position_ids or reset_attention_mask:
|
||||||
|
# Loop through the batches:
|
||||||
|
for b in range(micro_batch_size):
|
||||||
|
|
||||||
|
# Find indecies where EOD token is.
|
||||||
|
eod_index = position_ids[b, data[b] == eod_token]
|
||||||
|
# Detach indecies from positions if going to modify positions.
|
||||||
|
if reset_position_ids:
|
||||||
|
eod_index = eod_index.clone()
|
||||||
|
|
||||||
|
# Loop through EOD indecies:
|
||||||
|
prev_index = 0
|
||||||
|
for j in range(eod_index.size()[0]):
|
||||||
|
i = eod_index[j]
|
||||||
|
# Mask attention loss.
|
||||||
|
if reset_attention_mask:
|
||||||
|
attention_mask[b, 0, (i + 1) :, : (i + 1)] = 0
|
||||||
|
# Reset positions.
|
||||||
|
if reset_position_ids:
|
||||||
|
position_ids[b, (i + 1) :] -= i + 1 - prev_index
|
||||||
|
prev_index = i + 1
|
||||||
|
|
||||||
|
# Convert attention mask to binary:
|
||||||
|
attention_mask = attention_mask < 0.5
|
||||||
|
|
||||||
|
return attention_mask, loss_mask, position_ids
|
||||||
|
|
||||||
|
|
||||||
|
def get_batch(context_tokens: torch.LongTensor, eod_id: int):
|
||||||
|
"""Generate batch from context tokens."""
|
||||||
|
# Move to GPU.
|
||||||
|
tokens = context_tokens.contiguous().to(context_tokens.device)
|
||||||
|
# Get the attention mask and postition ids.
|
||||||
|
attention_mask, _, position_ids = get_ltor_masks_and_position_ids(
|
||||||
|
tokens,
|
||||||
|
eod_id,
|
||||||
|
reset_position_ids=False,
|
||||||
|
reset_attention_mask=False,
|
||||||
|
eod_mask_loss=False,
|
||||||
|
)
|
||||||
|
return tokens, attention_mask, position_ids
|
||||||
|
|
||||||
|
|
||||||
|
def get_stop_words_ids(chat_format, tokenizer):
|
||||||
|
if chat_format == "raw":
|
||||||
|
stop_words_ids = [tokenizer.encode("Human:"), [tokenizer.eod_id]]
|
||||||
|
elif chat_format == "chatml":
|
||||||
|
stop_words_ids = [[tokenizer.im_end_id], [tokenizer.im_start_id]]
|
||||||
|
else:
|
||||||
|
raise NotImplementedError(f"Unknown chat format {chat_format!r}")
|
||||||
|
return stop_words_ids
|
||||||
|
|
||||||
|
|
||||||
|
def make_context(
|
||||||
|
tokenizer: PreTrainedTokenizer,
|
||||||
|
query: str,
|
||||||
|
history: List[Tuple[str, str]] = None,
|
||||||
|
system: str = "",
|
||||||
|
max_window_size: int = 6144,
|
||||||
|
chat_format: str = "chatml",
|
||||||
|
):
|
||||||
|
if history is None:
|
||||||
|
history = []
|
||||||
|
|
||||||
|
if chat_format == "chatml":
|
||||||
|
im_start, im_end = "<|im_start|>", "<|im_end|>"
|
||||||
|
im_start_tokens = [tokenizer.im_start_id]
|
||||||
|
im_end_tokens = [tokenizer.im_end_id]
|
||||||
|
nl_tokens = tokenizer.encode("\n")
|
||||||
|
|
||||||
|
def _tokenize_str(role, content):
|
||||||
|
return f"{role}\n{content}", tokenizer.encode(
|
||||||
|
role, allowed_special=set()
|
||||||
|
) + nl_tokens + tokenizer.encode(content, allowed_special=set())
|
||||||
|
|
||||||
|
system_text, system_tokens_part = _tokenize_str("system", system)
|
||||||
|
system_tokens = im_start_tokens + system_tokens_part + im_end_tokens
|
||||||
|
|
||||||
|
raw_text = ""
|
||||||
|
context_tokens = []
|
||||||
|
|
||||||
|
for turn_query, turn_response in reversed(history):
|
||||||
|
query_text, query_tokens_part = _tokenize_str("user", turn_query)
|
||||||
|
query_tokens = im_start_tokens + query_tokens_part + im_end_tokens
|
||||||
|
response_text, response_tokens_part = _tokenize_str(
|
||||||
|
"assistant", turn_response
|
||||||
|
)
|
||||||
|
response_tokens = im_start_tokens + response_tokens_part + im_end_tokens
|
||||||
|
|
||||||
|
next_context_tokens = nl_tokens + query_tokens + nl_tokens + response_tokens
|
||||||
|
prev_chat = (
|
||||||
|
f"\n{im_start}{query_text}{im_end}\n{im_start}{response_text}{im_end}"
|
||||||
|
)
|
||||||
|
|
||||||
|
current_context_size = (
|
||||||
|
len(system_tokens) + len(next_context_tokens) + len(context_tokens)
|
||||||
|
)
|
||||||
|
if current_context_size < max_window_size:
|
||||||
|
context_tokens = next_context_tokens + context_tokens
|
||||||
|
raw_text = prev_chat + raw_text
|
||||||
|
else:
|
||||||
|
break
|
||||||
|
|
||||||
|
context_tokens = system_tokens + context_tokens
|
||||||
|
raw_text = f"{im_start}{system_text}{im_end}" + raw_text
|
||||||
|
context_tokens += (
|
||||||
|
nl_tokens
|
||||||
|
+ im_start_tokens
|
||||||
|
+ _tokenize_str("user", query)[1]
|
||||||
|
+ im_end_tokens
|
||||||
|
+ nl_tokens
|
||||||
|
+ im_start_tokens
|
||||||
|
+ tokenizer.encode("assistant")
|
||||||
|
+ nl_tokens
|
||||||
|
)
|
||||||
|
raw_text += f"\n{im_start}user\n{query}{im_end}\n{im_start}assistant\n"
|
||||||
|
|
||||||
|
elif chat_format == "raw":
|
||||||
|
raw_text = query
|
||||||
|
context_tokens = tokenizer.encode(raw_text)
|
||||||
|
else:
|
||||||
|
raise NotImplementedError(f"Unknown chat format {chat_format!r}")
|
||||||
|
|
||||||
|
return raw_text, context_tokens
|
||||||
|
|
||||||
|
|
||||||
|
def _decode_default(
|
||||||
|
tokens: List[int],
|
||||||
|
*,
|
||||||
|
stop_words: List[str],
|
||||||
|
eod_words: List[str],
|
||||||
|
tokenizer: PreTrainedTokenizer,
|
||||||
|
raw_text_len: int,
|
||||||
|
verbose: bool = False,
|
||||||
|
return_end_reason: bool = False,
|
||||||
|
errors: str='replace',
|
||||||
|
):
|
||||||
|
trim_decode_tokens = tokenizer.decode(tokens, errors=errors)[raw_text_len:]
|
||||||
|
if verbose:
|
||||||
|
print("\nRaw Generate: ", trim_decode_tokens)
|
||||||
|
|
||||||
|
end_reason = f"Gen length {len(tokens)}"
|
||||||
|
for stop_word in stop_words:
|
||||||
|
trim_decode_tokens = trim_decode_tokens.replace(stop_word, "").strip()
|
||||||
|
for eod_word in eod_words:
|
||||||
|
if eod_word in trim_decode_tokens:
|
||||||
|
end_reason = f"Gen {eod_word!r}"
|
||||||
|
trim_decode_tokens = trim_decode_tokens.split(eod_word)[0]
|
||||||
|
trim_decode_tokens = trim_decode_tokens.strip()
|
||||||
|
if verbose:
|
||||||
|
print("\nEnd Reason:", end_reason)
|
||||||
|
print("\nGenerate: ", trim_decode_tokens)
|
||||||
|
|
||||||
|
if return_end_reason:
|
||||||
|
return trim_decode_tokens, end_reason
|
||||||
|
else:
|
||||||
|
return trim_decode_tokens
|
||||||
|
|
||||||
|
|
||||||
|
def _decode_chatml(
|
||||||
|
tokens: List[int],
|
||||||
|
*,
|
||||||
|
stop_words: List[str],
|
||||||
|
eod_token_ids: List[int],
|
||||||
|
tokenizer: PreTrainedTokenizer,
|
||||||
|
raw_text_len: int,
|
||||||
|
context_length: int,
|
||||||
|
verbose: bool = False,
|
||||||
|
return_end_reason: bool = False,
|
||||||
|
errors: str='replace'
|
||||||
|
):
|
||||||
|
end_reason = f"Gen length {len(tokens)}"
|
||||||
|
eod_token_idx = context_length
|
||||||
|
for eod_token_idx in range(context_length, len(tokens)):
|
||||||
|
if tokens[eod_token_idx] in eod_token_ids:
|
||||||
|
end_reason = f"Gen {tokenizer.decode([tokens[eod_token_idx]])!r}"
|
||||||
|
break
|
||||||
|
|
||||||
|
trim_decode_tokens = tokenizer.decode(tokens[:eod_token_idx], errors=errors)[raw_text_len:]
|
||||||
|
if verbose:
|
||||||
|
print("\nRaw Generate w/o EOD:", tokenizer.decode(tokens, errors=errors)[raw_text_len:])
|
||||||
|
print("\nRaw Generate:", trim_decode_tokens)
|
||||||
|
print("\nEnd Reason:", end_reason)
|
||||||
|
for stop_word in stop_words:
|
||||||
|
trim_decode_tokens = trim_decode_tokens.replace(stop_word, "").strip()
|
||||||
|
trim_decode_tokens = trim_decode_tokens.strip()
|
||||||
|
if verbose:
|
||||||
|
print("\nGenerate:", trim_decode_tokens)
|
||||||
|
|
||||||
|
if return_end_reason:
|
||||||
|
return trim_decode_tokens, end_reason
|
||||||
|
else:
|
||||||
|
return trim_decode_tokens
|
||||||
|
|
||||||
|
|
||||||
|
def decode_tokens(
|
||||||
|
tokens: Union[torch.LongTensor, TokensType],
|
||||||
|
tokenizer: PreTrainedTokenizer,
|
||||||
|
raw_text_len: int,
|
||||||
|
context_length: int,
|
||||||
|
chat_format: str,
|
||||||
|
verbose: bool = False,
|
||||||
|
return_end_reason: bool = False,
|
||||||
|
errors: str="replace",
|
||||||
|
) -> str:
|
||||||
|
if torch.is_tensor(tokens):
|
||||||
|
tokens = tokens.cpu().numpy().tolist()
|
||||||
|
|
||||||
|
if chat_format == "chatml":
|
||||||
|
return _decode_chatml(
|
||||||
|
tokens,
|
||||||
|
stop_words=[],
|
||||||
|
eod_token_ids=[tokenizer.im_start_id, tokenizer.im_end_id],
|
||||||
|
tokenizer=tokenizer,
|
||||||
|
raw_text_len=raw_text_len,
|
||||||
|
context_length=context_length,
|
||||||
|
verbose=verbose,
|
||||||
|
return_end_reason=return_end_reason,
|
||||||
|
errors=errors,
|
||||||
|
)
|
||||||
|
elif chat_format == "raw":
|
||||||
|
return _decode_default(
|
||||||
|
tokens,
|
||||||
|
stop_words=["<|endoftext|>"],
|
||||||
|
eod_words=["<|endoftext|>"],
|
||||||
|
tokenizer=tokenizer,
|
||||||
|
raw_text_len=raw_text_len,
|
||||||
|
verbose=verbose,
|
||||||
|
return_end_reason=return_end_reason,
|
||||||
|
errors=errors,
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
raise NotImplementedError(f"Unknown chat format {chat_format!r}")
|
||||||
|
|
||||||
|
|
||||||
|
class StopWordsLogitsProcessor(LogitsProcessor):
|
||||||
|
"""
|
||||||
|
:class:`transformers.LogitsProcessor` that enforces that when specified sequences appear, stop geration.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
stop_words_ids (:obj:`List[List[int]]`):
|
||||||
|
List of list of token ids of stop ids. In order to get the tokens of the words
|
||||||
|
that should not appear in the generated text, use :obj:`tokenizer(bad_word,
|
||||||
|
add_prefix_space=True).input_ids`.
|
||||||
|
eos_token_id (:obj:`int`):
|
||||||
|
The id of the `end-of-sequence` token.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, stop_words_ids: Iterable[Iterable[int]], eos_token_id: int):
|
||||||
|
|
||||||
|
if not isinstance(stop_words_ids, List) or len(stop_words_ids) == 0:
|
||||||
|
raise ValueError(
|
||||||
|
f"`stop_words_ids` has to be a non-emtpy list, but is {stop_words_ids}."
|
||||||
|
)
|
||||||
|
if any(not isinstance(bad_word_ids, list) for bad_word_ids in stop_words_ids):
|
||||||
|
raise ValueError(
|
||||||
|
f"`stop_words_ids` has to be a list of lists, but is {stop_words_ids}."
|
||||||
|
)
|
||||||
|
if any(
|
||||||
|
any(
|
||||||
|
(not isinstance(token_id, (int, np.integer)) or token_id < 0)
|
||||||
|
for token_id in stop_word_ids
|
||||||
|
)
|
||||||
|
for stop_word_ids in stop_words_ids
|
||||||
|
):
|
||||||
|
raise ValueError(
|
||||||
|
f"Each list in `stop_words_ids` has to be a list of positive integers, but is {stop_words_ids}."
|
||||||
|
)
|
||||||
|
|
||||||
|
self.stop_words_ids = list(
|
||||||
|
filter(
|
||||||
|
lambda bad_token_seq: bad_token_seq != [eos_token_id], stop_words_ids
|
||||||
|
)
|
||||||
|
)
|
||||||
|
self.eos_token_id = eos_token_id
|
||||||
|
for stop_token_seq in self.stop_words_ids:
|
||||||
|
assert (
|
||||||
|
len(stop_token_seq) > 0
|
||||||
|
), "Stop words token sequences {} cannot have an empty list".format(
|
||||||
|
stop_words_ids
|
||||||
|
)
|
||||||
|
|
||||||
|
def __call__(
|
||||||
|
self, input_ids: torch.LongTensor, scores: torch.FloatTensor
|
||||||
|
) -> torch.FloatTensor:
|
||||||
|
stopped_samples = self._calc_stopped_samples(input_ids)
|
||||||
|
for i, should_stop in enumerate(stopped_samples):
|
||||||
|
if should_stop:
|
||||||
|
scores[i, self.eos_token_id] = float(2**15)
|
||||||
|
return scores
|
||||||
|
|
||||||
|
def _tokens_match(self, prev_tokens: torch.LongTensor, tokens: List[int]) -> bool:
|
||||||
|
if len(tokens) == 0:
|
||||||
|
# if bad word tokens is just one token always ban it
|
||||||
|
return True
|
||||||
|
elif len(tokens) > len(prev_tokens):
|
||||||
|
# if bad word tokens are longer then prev input_ids they can't be equal
|
||||||
|
return False
|
||||||
|
elif prev_tokens[-len(tokens) :].tolist() == tokens:
|
||||||
|
# if tokens match
|
||||||
|
return True
|
||||||
|
else:
|
||||||
|
return False
|
||||||
|
|
||||||
|
def _calc_stopped_samples(self, prev_input_ids: Iterable[int]) -> Iterable[int]:
|
||||||
|
stopped_samples = []
|
||||||
|
for prev_input_ids_slice in prev_input_ids:
|
||||||
|
match = False
|
||||||
|
for stop_token_seq in self.stop_words_ids:
|
||||||
|
if self._tokens_match(prev_input_ids_slice, stop_token_seq):
|
||||||
|
# if tokens do not match continue
|
||||||
|
match = True
|
||||||
|
break
|
||||||
|
stopped_samples.append(match)
|
||||||
|
|
||||||
|
return stopped_samples
|
||||||
|
|
||||||
|
|
||||||
|
def top_k_logits(logits, top_k=0, top_p=0.0, filter_value=-float("Inf")):
|
||||||
|
"""This function has been mostly taken from huggingface conversational
|
||||||
|
ai code at
|
||||||
|
https://medium.com/huggingface/how-to-build-a-state-of-the-art-
|
||||||
|
conversational-ai-with-transfer-learning-2d818ac26313"""
|
||||||
|
|
||||||
|
if top_k > 0:
|
||||||
|
# Remove all tokens with a probability less than the
|
||||||
|
# last token of the top-k
|
||||||
|
indices_to_remove = logits < torch.topk(logits, top_k)[0][..., -1, None]
|
||||||
|
logits[indices_to_remove] = filter_value
|
||||||
|
|
||||||
|
if top_p > 0.0:
|
||||||
|
# Cconvert to 1D
|
||||||
|
sorted_logits, sorted_indices = torch.sort(logits, descending=True, dim=-1)
|
||||||
|
cumulative_probs = torch.cumsum(F.softmax(sorted_logits, dim=-1), dim=-1)
|
||||||
|
|
||||||
|
# Remove tokens with cumulative probability above the threshold
|
||||||
|
sorted_indices_to_remove = cumulative_probs > top_p
|
||||||
|
# Shift the indices to the right to keep also the first token
|
||||||
|
# above the threshold
|
||||||
|
sorted_indices_to_remove[..., 1:] = sorted_indices_to_remove[..., :-1].clone()
|
||||||
|
sorted_indices_to_remove[..., 0] = 0
|
||||||
|
for i in range(sorted_indices.size(0)):
|
||||||
|
indices_to_remove = sorted_indices[i][sorted_indices_to_remove[i]]
|
||||||
|
logits[i][indices_to_remove] = filter_value
|
||||||
|
|
||||||
|
return logits
|
||||||
|
|
||||||
|
|
||||||
|
def switch(val1, val2, boolean):
|
||||||
|
boolean = boolean.type_as(val1)
|
||||||
|
return (1 - boolean) * val1 + boolean * val2
|
|
@ -0,0 +1,276 @@
|
||||||
|
# Copyright (c) Alibaba Cloud.
|
||||||
|
#
|
||||||
|
# This source code is licensed under the license found in the
|
||||||
|
# LICENSE file in the root directory of this source tree.
|
||||||
|
|
||||||
|
"""Tokenization classes for QWen."""
|
||||||
|
|
||||||
|
import base64
|
||||||
|
import logging
|
||||||
|
import os
|
||||||
|
import unicodedata
|
||||||
|
from typing import Collection, Dict, List, Set, Tuple, Union
|
||||||
|
|
||||||
|
import tiktoken
|
||||||
|
from transformers import PreTrainedTokenizer, AddedToken
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
VOCAB_FILES_NAMES = {"vocab_file": "qwen.tiktoken"}
|
||||||
|
|
||||||
|
PAT_STR = r"""(?i:'s|'t|'re|'ve|'m|'ll|'d)|[^\r\n\p{L}\p{N}]?\p{L}+|\p{N}| ?[^\s\p{L}\p{N}]+[\r\n]*|\s*[\r\n]+|\s+(?!\S)|\s+"""
|
||||||
|
ENDOFTEXT = "<|endoftext|>"
|
||||||
|
IMSTART = "<|im_start|>"
|
||||||
|
IMEND = "<|im_end|>"
|
||||||
|
# as the default behavior is changed to allow special tokens in
|
||||||
|
# regular texts, the surface forms of special tokens need to be
|
||||||
|
# as different as possible to minimize the impact
|
||||||
|
EXTRAS = tuple((f"<|extra_{i}|>" for i in range(205)))
|
||||||
|
# changed to use actual index to avoid misconfiguration with vocabulary expansion
|
||||||
|
SPECIAL_START_ID = 151643
|
||||||
|
SPECIAL_TOKENS = tuple(
|
||||||
|
enumerate(
|
||||||
|
(
|
||||||
|
(
|
||||||
|
ENDOFTEXT,
|
||||||
|
IMSTART,
|
||||||
|
IMEND,
|
||||||
|
)
|
||||||
|
+ EXTRAS
|
||||||
|
),
|
||||||
|
start=SPECIAL_START_ID,
|
||||||
|
)
|
||||||
|
)
|
||||||
|
SPECIAL_TOKENS_SET = set(t for i, t in SPECIAL_TOKENS)
|
||||||
|
|
||||||
|
|
||||||
|
def _load_tiktoken_bpe(tiktoken_bpe_file: str) -> Dict[bytes, int]:
|
||||||
|
with open(tiktoken_bpe_file, "rb") as f:
|
||||||
|
contents = f.read()
|
||||||
|
return {
|
||||||
|
base64.b64decode(token): int(rank)
|
||||||
|
for token, rank in (line.split() for line in contents.splitlines() if line)
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class QWenTokenizer(PreTrainedTokenizer):
|
||||||
|
"""QWen tokenizer."""
|
||||||
|
|
||||||
|
vocab_files_names = VOCAB_FILES_NAMES
|
||||||
|
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
vocab_file,
|
||||||
|
errors="replace",
|
||||||
|
extra_vocab_file=None,
|
||||||
|
**kwargs,
|
||||||
|
):
|
||||||
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
|
# how to handle errors in decoding UTF-8 byte sequences
|
||||||
|
# use ignore if you are in streaming inference
|
||||||
|
self.errors = errors
|
||||||
|
|
||||||
|
self.mergeable_ranks = _load_tiktoken_bpe(vocab_file) # type: Dict[bytes, int]
|
||||||
|
self.special_tokens = {
|
||||||
|
token: index
|
||||||
|
for index, token in SPECIAL_TOKENS
|
||||||
|
}
|
||||||
|
|
||||||
|
# try load extra vocab from file
|
||||||
|
if extra_vocab_file is not None:
|
||||||
|
used_ids = set(self.mergeable_ranks.values()) | set(self.special_tokens.values())
|
||||||
|
extra_mergeable_ranks = _load_tiktoken_bpe(extra_vocab_file)
|
||||||
|
for token, index in extra_mergeable_ranks.items():
|
||||||
|
if token in self.mergeable_ranks:
|
||||||
|
logger.info(f"extra token {token} exists, skipping")
|
||||||
|
continue
|
||||||
|
if index in used_ids:
|
||||||
|
logger.info(f'the index {index} for extra token {token} exists, skipping')
|
||||||
|
continue
|
||||||
|
self.mergeable_ranks[token] = index
|
||||||
|
# the index may be sparse after this, but don't worry tiktoken.Encoding will handle this
|
||||||
|
|
||||||
|
enc = tiktoken.Encoding(
|
||||||
|
"Qwen",
|
||||||
|
pat_str=PAT_STR,
|
||||||
|
mergeable_ranks=self.mergeable_ranks,
|
||||||
|
special_tokens=self.special_tokens,
|
||||||
|
)
|
||||||
|
assert (
|
||||||
|
len(self.mergeable_ranks) + len(self.special_tokens) == enc.n_vocab
|
||||||
|
), f"{len(self.mergeable_ranks) + len(self.special_tokens)} != {enc.n_vocab} in encoding"
|
||||||
|
|
||||||
|
self.decoder = {
|
||||||
|
v: k for k, v in self.mergeable_ranks.items()
|
||||||
|
} # type: dict[int, bytes|str]
|
||||||
|
self.decoder.update({v: k for k, v in self.special_tokens.items()})
|
||||||
|
|
||||||
|
self.tokenizer = enc # type: tiktoken.Encoding
|
||||||
|
|
||||||
|
self.eod_id = self.tokenizer.eot_token
|
||||||
|
self.im_start_id = self.special_tokens[IMSTART]
|
||||||
|
self.im_end_id = self.special_tokens[IMEND]
|
||||||
|
|
||||||
|
def __getstate__(self):
|
||||||
|
# for pickle lovers
|
||||||
|
state = self.__dict__.copy()
|
||||||
|
del state["tokenizer"]
|
||||||
|
return state
|
||||||
|
|
||||||
|
def __setstate__(self, state):
|
||||||
|
# tokenizer is not python native; don't pass it; rebuild it
|
||||||
|
self.__dict__.update(state)
|
||||||
|
enc = tiktoken.Encoding(
|
||||||
|
"Qwen",
|
||||||
|
pat_str=PAT_STR,
|
||||||
|
mergeable_ranks=self.mergeable_ranks,
|
||||||
|
special_tokens=self.special_tokens,
|
||||||
|
)
|
||||||
|
self.tokenizer = enc
|
||||||
|
|
||||||
|
def __len__(self) -> int:
|
||||||
|
return self.tokenizer.n_vocab
|
||||||
|
|
||||||
|
def get_vocab(self) -> Dict[bytes, int]:
|
||||||
|
return self.mergeable_ranks
|
||||||
|
|
||||||
|
def convert_tokens_to_ids(
|
||||||
|
self, tokens: Union[bytes, str, List[Union[bytes, str]]]
|
||||||
|
) -> List[int]:
|
||||||
|
ids = []
|
||||||
|
if isinstance(tokens, (str, bytes)):
|
||||||
|
if tokens in self.special_tokens:
|
||||||
|
return self.special_tokens[tokens]
|
||||||
|
else:
|
||||||
|
return self.mergeable_ranks.get(tokens)
|
||||||
|
for token in tokens:
|
||||||
|
if token in self.special_tokens:
|
||||||
|
ids.append(self.special_tokens[token])
|
||||||
|
else:
|
||||||
|
ids.append(self.mergeable_ranks.get(token))
|
||||||
|
return ids
|
||||||
|
|
||||||
|
def _add_tokens(
|
||||||
|
self,
|
||||||
|
new_tokens: Union[List[str], List[AddedToken]],
|
||||||
|
special_tokens: bool = False,
|
||||||
|
) -> int:
|
||||||
|
if not special_tokens and new_tokens:
|
||||||
|
raise ValueError("Adding regular tokens is not supported")
|
||||||
|
for token in new_tokens:
|
||||||
|
surface_form = token.content if isinstance(token, AddedToken) else token
|
||||||
|
if surface_form not in SPECIAL_TOKENS_SET:
|
||||||
|
raise ValueError("Adding unknown special tokens is not supported")
|
||||||
|
return 0
|
||||||
|
|
||||||
|
def save_vocabulary(self, save_directory: str, **kwargs) -> Tuple[str]:
|
||||||
|
"""
|
||||||
|
Save only the vocabulary of the tokenizer (vocabulary).
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
`Tuple(str)`: Paths to the files saved.
|
||||||
|
"""
|
||||||
|
file_path = os.path.join(save_directory, "qwen.tiktoken")
|
||||||
|
with open(file_path, "w", encoding="utf8") as w:
|
||||||
|
for k, v in self.mergeable_ranks.items():
|
||||||
|
line = base64.b64encode(k).decode("utf8") + " " + str(v) + "\n"
|
||||||
|
w.write(line)
|
||||||
|
return (file_path,)
|
||||||
|
|
||||||
|
def tokenize(
|
||||||
|
self,
|
||||||
|
text: str,
|
||||||
|
allowed_special: Union[Set, str] = "all",
|
||||||
|
disallowed_special: Union[Collection, str] = (),
|
||||||
|
**kwargs,
|
||||||
|
) -> List[Union[bytes, str]]:
|
||||||
|
"""
|
||||||
|
Converts a string in a sequence of tokens.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
text (`str`):
|
||||||
|
The sequence to be encoded.
|
||||||
|
allowed_special (`Literal["all"]` or `set`):
|
||||||
|
The surface forms of the tokens to be encoded as special tokens in regular texts.
|
||||||
|
Default to "all".
|
||||||
|
disallowed_special (`Literal["all"]` or `Collection`):
|
||||||
|
The surface forms of the tokens that should not be in regular texts and trigger errors.
|
||||||
|
Default to an empty tuple.
|
||||||
|
|
||||||
|
kwargs (additional keyword arguments, *optional*):
|
||||||
|
Will be passed to the underlying model specific encode method.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
`List[bytes|str]`: The list of tokens.
|
||||||
|
"""
|
||||||
|
tokens = []
|
||||||
|
text = unicodedata.normalize("NFC", text)
|
||||||
|
|
||||||
|
# this implementation takes a detour: text -> token id -> token surface forms
|
||||||
|
for t in self.tokenizer.encode(
|
||||||
|
text, allowed_special=allowed_special, disallowed_special=disallowed_special
|
||||||
|
):
|
||||||
|
tokens.append(self.decoder[t])
|
||||||
|
return tokens
|
||||||
|
|
||||||
|
def convert_tokens_to_string(self, tokens: List[Union[bytes, str]]) -> str:
|
||||||
|
"""
|
||||||
|
Converts a sequence of tokens in a single string.
|
||||||
|
"""
|
||||||
|
text = ""
|
||||||
|
temp = b""
|
||||||
|
for t in tokens:
|
||||||
|
if isinstance(t, str):
|
||||||
|
if temp:
|
||||||
|
text += temp.decode("utf-8", errors=self.errors)
|
||||||
|
temp = b""
|
||||||
|
text += t
|
||||||
|
elif isinstance(t, bytes):
|
||||||
|
temp += t
|
||||||
|
else:
|
||||||
|
raise TypeError("token should only be of type types or str")
|
||||||
|
if temp:
|
||||||
|
text += temp.decode("utf-8", errors=self.errors)
|
||||||
|
return text
|
||||||
|
|
||||||
|
@property
|
||||||
|
def vocab_size(self):
|
||||||
|
return self.tokenizer.n_vocab
|
||||||
|
|
||||||
|
def _convert_id_to_token(self, index: int) -> Union[bytes, str]:
|
||||||
|
"""Converts an id to a token, special tokens included"""
|
||||||
|
if index in self.decoder:
|
||||||
|
return self.decoder[index]
|
||||||
|
raise ValueError("unknown ids")
|
||||||
|
|
||||||
|
def _convert_token_to_id(self, token: Union[bytes, str]) -> int:
|
||||||
|
"""Converts a token to an id using the vocab, special tokens included"""
|
||||||
|
if token in self.special_tokens:
|
||||||
|
return self.special_tokens[token]
|
||||||
|
if token in self.mergeable_ranks:
|
||||||
|
return self.mergeable_ranks[token]
|
||||||
|
raise ValueError("unknown token")
|
||||||
|
|
||||||
|
def _tokenize(self, text: str, **kwargs):
|
||||||
|
"""
|
||||||
|
Converts a string in a sequence of tokens (string), using the tokenizer. Split in words for word-based
|
||||||
|
vocabulary or sub-words for sub-word-based vocabularies (BPE/SentencePieces/WordPieces).
|
||||||
|
|
||||||
|
Do NOT take care of added tokens.
|
||||||
|
"""
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
|
def _decode(
|
||||||
|
self,
|
||||||
|
token_ids: Union[int, List[int]],
|
||||||
|
skip_special_tokens: bool = False,
|
||||||
|
errors: str = None,
|
||||||
|
**kwargs,
|
||||||
|
) -> str:
|
||||||
|
if isinstance(token_ids, int):
|
||||||
|
token_ids = [token_ids]
|
||||||
|
if skip_special_tokens:
|
||||||
|
token_ids = [i for i in token_ids if i < self.eod_id]
|
||||||
|
return self.tokenizer.decode(token_ids, errors=errors or self.errors)
|
|
@ -0,0 +1,10 @@
|
||||||
|
{
|
||||||
|
"model_max_length": 8192,
|
||||||
|
"tokenizer_class": "QWenTokenizer",
|
||||||
|
"auto_map": {
|
||||||
|
"AutoTokenizer": [
|
||||||
|
"tokenization_qwen.QWenTokenizer",
|
||||||
|
null
|
||||||
|
]
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue