Update code.

This commit is contained in:
Colin 2023-12-22 18:01:57 +08:00
parent 185caa12e9
commit 10268c4414
4 changed files with 77 additions and 52 deletions

23
Readme.md Normal file
View File

@ -0,0 +1,23 @@
## data flow
input_ids = tokenizer.build_chat_input(query, history=history, role=role)
input_ids -> [1, 6]
inputs_embeds -> [6, 1, 4096] 4096:hidden_size
rotary_pos_emb -> [6, 1, 32, 2] 32:pos的编码维度 2:cos+sin
hidden_states = inputs_embeds
for layers : GLMBlock(hidden_states, rotary_pos_emb)
hidden_states = self.final_layernorm(hidden_states)
hidden_states = hidden_states[-1:]
lm_logits = self.output_layer(hidden_states)
lm_logits = lm_logits.transpose(0, 1).contiguous() -> [1, 1, 65024]
probs = softmax(lm_logits) -> [1, 65024]
next_tokens = torch.multinomial(probs, num_samples=1) 采样 -> [1]
input_ids = torch.cat([input_ids, next_tokens) -> [1, 7]
response = tokenizer.decode(outputs)

View File

@ -170,7 +170,7 @@ class SelfAttention(torch.nn.Module):
x_out2 = x_out2.flatten(3)
return torch.cat((x_out2, x_pass), dim=-1)
def forward(self, hidden_states, rotary_pos_emb, kv_cache=None):
def forward(self, hidden_states, rotary_pos_emb):
# hidden_states: [sq, b, h]
# Attention heads [sq, b, h] --> [sq, b, (np * 3 * hn)]
mixed_x_layer = self.query_key_value(hidden_states)
@ -213,8 +213,6 @@ class SelfAttention(torch.nn.Module):
query_layer = self.apply_rotary_pos_emb(query_layer, rotary_pos_emb)
key_layer = self.apply_rotary_pos_emb(key_layer, rotary_pos_emb)
kv_cache = (key_layer, value_layer)
key_layer = key_layer.unsqueeze(-2)
key_layer = key_layer.expand(
-1,
@ -255,7 +253,7 @@ class SelfAttention(torch.nn.Module):
# Output. [sq, b, h]
# =================
output = self.dense(context_layer)
return output, kv_cache
return output
class MLP(torch.nn.Module):
@ -342,14 +340,12 @@ class GLMBlock(torch.nn.Module):
# MLP
self.mlp = MLP(config, device=device)
def forward(self, hidden_states, rotary_pos_emb, kv_cache=None):
def forward(self, hidden_states, rotary_pos_emb):
# hidden_states: [s, b, h]
# Layer norm at the beginning of the transformer layer.
layernorm_output = self.input_layernorm(hidden_states)
# Self attention.
attention_output, kv_cache = self.self_attention(
layernorm_output, rotary_pos_emb, kv_cache=kv_cache
)
attention_output = self.self_attention(layernorm_output, rotary_pos_emb)
residual = hidden_states
layernorm_input = torch.nn.functional.dropout(
@ -369,7 +365,7 @@ class GLMBlock(torch.nn.Module):
mlp_output, p=self.hidden_dropout, training=self.training
)
output = residual + output
return output, kv_cache
return output
class GLMTransformer(torch.nn.Module):
@ -389,18 +385,10 @@ class GLMTransformer(torch.nn.Module):
dtype=config.torch_dtype,
)
def forward(
self,
hidden_states,
rotary_pos_emb
):
kv_caches = [None for _ in range(self.num_layers)]
def forward(self, hidden_states, rotary_pos_emb):
for index in range(self.num_layers):
layer = self.layers[index]
hidden_states, kv_cache = layer(
hidden_states, rotary_pos_emb, kv_cache=kv_caches[index]
)
hidden_states = layer(hidden_states, rotary_pos_emb)
hidden_states = self.final_layernorm(hidden_states)
return hidden_states
@ -469,27 +457,20 @@ class ChatGLMModel(nn.Module):
input_ids,
position_ids: Optional[torch.Tensor] = None,
output_hidden_states: Optional[bool] = None,
return_last_logit: Optional[bool] = False,
):
output_hidden_states = (
output_hidden_states
if output_hidden_states is not None
else self.config.output_hidden_states
)
batch_size, seq_length = input_ids.shape
inputs_embeds = self.embedding(input_ids)
# Rotary positional embeddings
rotary_pos_emb = self.rotary_pos_emb(self.seq_length)
# show.DumpTensorToImage(rotary_pos_emb[:, :, 0], "rotary_pos_emb.png", scale=0.1)
rotary_pos_emb = rotary_pos_emb[position_ids]
rotary_pos_emb = rotary_pos_emb.transpose(0, 1).contiguous()
hidden_states = self.encoder(
inputs_embeds,
rotary_pos_emb=rotary_pos_emb
)
if return_last_logit:
hidden_states = self.encoder(inputs_embeds, rotary_pos_emb)
hidden_states = hidden_states[-1:]
lm_logits = self.output_layer(hidden_states)
lm_logits = lm_logits.transpose(0, 1).contiguous()
@ -676,7 +657,7 @@ class ChatGLMForConditionalGeneration(nn.Module):
input_ids,
pad_token_id=generation_config.pad_token_id,
eos_token_id=generation_config.eos_token_id,
output_hidden_states=generation_config.output_hidden_states
output_hidden_states=generation_config.output_hidden_states,
)
outputs = outputs.tolist()[0][len(inputs["input_ids"][0]) : -1]
@ -689,7 +670,7 @@ class ChatGLMForConditionalGeneration(nn.Module):
input_ids: torch.LongTensor,
pad_token_id: Optional[int] = None,
eos_token_id: Optional[Union[int, List[int]]] = None,
output_hidden_states: Optional[bool] = None
output_hidden_states: Optional[bool] = None,
):
if isinstance(eos_token_id, int):
eos_token_id = [eos_token_id]
@ -708,11 +689,7 @@ class ChatGLMForConditionalGeneration(nn.Module):
.unsqueeze(0)
.repeat(batch_size, 1)
)
model_inputs = {
"input_ids": input_ids_in,
"position_ids": position_ids_in,
"return_last_logit": True
}
model_inputs = {"input_ids": input_ids_in, "position_ids": position_ids_in}
logits = self.transformer(
**model_inputs,
@ -723,7 +700,6 @@ class ChatGLMForConditionalGeneration(nn.Module):
next_tokens = torch.multinomial(probs, num_samples=1).squeeze(1)
# finished sentences should have their next token be a padding token
if eos_token_id is not None:
next_tokens = next_tokens * unfinished_sequences + pad_token_id * (
1 - unfinished_sequences
)
@ -732,15 +708,12 @@ class ChatGLMForConditionalGeneration(nn.Module):
input_ids = torch.cat([input_ids, next_tokens[:, None]], dim=-1)
# if eos_token was found in one sentence, set sentence to finished
if eos_token_id_tensor is not None:
unfinished_sequences = unfinished_sequences.mul(
next_tokens.tile(eos_token_id_tensor.shape[0], 1)
.ne(eos_token_id_tensor.unsqueeze(1))
.prod(dim=0)
)
if unfinished_sequences.max() == 0:
this_peer_finished = True
if this_peer_finished:
break
return input_ids

View File

@ -25,7 +25,7 @@ if tokenizer_config_file is not None:
init_kwargs.pop("tokenizer_file", None)
saved_init_inputs = init_kwargs.pop("init_inputs", ())
init_inputs = saved_init_inputs
init_kwargs["vocab_file"] = './chatglm/tokenizer.model'
init_kwargs["vocab_file"] = "./chatglm/tokenizer.model"
init_kwargs["added_tokens_file"] = None
init_kwargs["special_tokens_map_file"] = None
init_kwargs["tokenizer_file"] = None
@ -35,9 +35,11 @@ tokenizer = ChatGLMTokenizer(*init_inputs, **init_kwargs)
glm = glm.from_pretrained(pretrained_model_name_or_path, config=config).half().cuda()
glm = glm.eval()
response, history = glm.chat(tokenizer, "colin", history=[])
query = "colin"
response, history = glm.chat(tokenizer, query, history=[])
print(response)
response, history = glm.chat(tokenizer, "你好", history=history)
query = "你好"
response, history = glm.chat(tokenizer, query, history=history)
print(response)
# response, history = glm.chat(tokenizer, "你是一个心理学专家,请问晚上睡不着应该怎么办", history=history)
# print(response)
@ -50,7 +52,6 @@ print(response)
# px.scatter(gapminder2007, x='gdpPercap', y='lifeExp')
# from modelscope import AutoTokenizer, AutoModel, snapshot_download
# model_dir = snapshot_download("ZhipuAI/chatglm3-6b", cache_dir="./chatglm", revision="v1.0.0")
# model = AutoModel.from_pretrained(model_dir, trust_remote_code=True).half().cuda()

28
embedding.py Normal file
View File

@ -0,0 +1,28 @@
import torch
import torch.nn as nn
# 定义词表大小和向量维度
vocab_size = 10000
embedding_dim = 16
# 定义一个Embedding层
embedding = nn.Embedding(num_embeddings=vocab_size, embedding_dim=embedding_dim)
# 定义一个输入张量,形状为(batch_size, sequence_length)
input_tensor = torch.LongTensor([[1, 2], [4, 3]])
# 将输入张量传递给Embedding层
embedded_tensor = embedding(input_tensor)
print("embedded weight shape:")
print(embedding.weight.shape)
print("embedded weight:")
print(embedding.weight)
# 输出形状为 (batch_size, sequence_length, embedding_dim)
print("embedded out shape:")
print(embedded_tensor.shape)
print("embedded out:")
print(embedded_tensor)