On this work, we deal with the challenge of zero-shot cross-domain DST through leveraging giant scale pre-trained sequence-to-sequence (seq2seq) fashions and with effective encoding of slot descriptions. In addition, we exhibit that joint training as well as the usage of pre-skilled language models, ELMo and BERT in our case, are complementary to those few-shot studying…