import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
from sklearn.model_selection import train_test_split
from keras.models import Sequential
from keras.layers import Dense, Flatten, Conv1D, MaxPooling1D, LSTM, Reshape
from keras.utils import to_categorical

def read_fasta(file):
    seqs = []
    labels = []
    with open(file, 'r') as f:
        for line in f:
            if line.startswith('>'):
                labels.append(line.strip()[1:])
            else:
                seqs.append(line.strip())
    label_dict = {l:i for i,l in enumerate(set(labels))}
    label_indices = [label_dict[l] for l in labels]
    seqs = np.array(seqs)[np.argsort(label_indices)]
    labels = np.array(label_indices)
    return seqs, labels

def seq_to_num(seqs):
    seq_num = []
    for seq in seqs:
        seq_num.append([int(i) for i in seq.replace('A', '1').replace('C', '2').replace('G', '3').replace('T', '4')])
    return np.array(seq_num)

def encode_labels(labels):
    label_dict = {l:i for i,l in enumerate(set(labels))}
    encoded_labels = [label_dict[l] for l in labels]
    return to_categorical(encoded_labels)

def create_model():
    model = Sequential()
    model.add(Conv1D(filters=32, kernel_size=3, activation='relu', input_shape=(1000, 1)))
    model.add(MaxPooling1D(pool_size=2))
    model.add(Conv1D(filters=64, kernel_size=3, activation='relu'))
    model.add(MaxPooling1D(pool_size=2))
    model.add(Flatten())
    model.add(Reshape((-1, 64)))  # 新增的Reshape层
    model.add(LSTM(64))
    model.add(Dense(64, activation='relu'))
    model.add(Dense(2, activation='softmax'))
    model.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])
    return model

seqs, labels = read_fasta('CP015726.fasta')
seq_num = seq_to_num(seqs)
encoded_labels = encode_labels(labels)
X_train, X_test, y_train, y_test = train_test_split(seq_num, encoded_labels, test_size=0.2, random_state=42)
model = create_model()
history = model.fit(X_train.reshape((-1, 1000, 1)), y_train, epochs=10, batch_size=32, validation_split=0.2)
score = model.evaluate(X_test.reshape((-1, 1000, 1)), y_test, verbose=0)
print('Test loss:', score[0])
print('Test accuracy:', score[1])

# ValueError: cannot reshape array of size 151 into shape (1000,1)
# 错误分析
# 从错误信息可以看出,无法将大小为151的数组重塑为(1000,1)的形状,因此需要检查数据的形状。
# X_train和X_test是序列数据,每个序列的长度是1000,因此它们的形状应该是(n_samples, 1000)。
# 但是,在传递给模型之前,需要将每个序列转换为1D张量,以便Conv1D层可以处理它们。因此,需要将X_train和X_test重塑为(n_samples, 1000, 1)的形状,表示每个序列有1000个特征,每个特征是一个标量。
# 在重塑之前,需要确保X_train和X_test中的所有序列都具有相同的长度,如果有任何序列长度不足1000,则需要进行填充。可以使用Keras的pad_sequences函数来完成填充。
DNA序列分类的深度学习模型:使用CNN-LSTM架构

原文地址: https://www.cveoy.top/t/topic/lKSB 著作权归作者所有。请勿转载和采集!

免费AI点我,无需注册和登录