Подписаться
Zixuan Ke
Zixuan Ke
Salesforce AI Research
Подтвержден адрес электронной почты в домене uic.edu - Главная страница
Название
Процитировано
Процитировано
Год
Automated Essay Scoring: A Survey of the State of the Art.
Z Ke, V Ng
IJCAI 19, 6300-6308, 2019
2702019
Continual learning of a mixed sequence of similar and dissimilar tasks
Z Ke, B Liu, X Huang
Advances in neural information processing systems 33, 18493-18504, 2020
1472020
Continual pre-training of language models
Z Ke, Y Shao, H Lin, T Konishi, G Kim, B Liu
The Eleventh International Conference on Learning Representations (ICLR 2023), 2023
1332023
Achieving forgetting prevention and knowledge transfer in continual learning
Z Ke, B Liu, N Ma, H Xu, L Shu
Advances in Neural Information Processing Systems 34, 22443-22456, 2021
1302021
Give me more feedback: Annotating argument persuasiveness and related attributes in student essays
W Carlile, N Gurrapadi, Z Ke, V Ng
Proceedings of the 56th Annual Meeting of the Association for Computational …, 2018
972018
A theoretical study on solving continual learning
G Kim, C Xiao, T Konishi, Z Ke, B Liu
Advances in neural information processing systems 35, 5065-5079, 2022
962022
Continual learning of natural language processing tasks: A survey
Z Ke, B Liu
arXiv preprint arXiv:2211.12701, 2022
882022
Adapting BERT for continual learning of a sequence of aspect sentiment classification tasks
Z Ke, H Xu, B Liu
Proceedings of the 2021 Conference of the North American Chapter of the …, 2021
832021
CLASSIC: Continual and contrastive learning of aspect sentiment classification tasks
Z Ke, B Liu, H Xu, L Shu
Proceedings of 2021 Conference on Empirical Methods in Natural Language …, 2021
642021
Learning to give feedback: Modeling attributes affecting argument persuasiveness in student essays.
Z Ke, W Carlile, N Gurrapadi, V Ng
IJCAI, 4130-4136, 2018
462018
A multi-head model for continual learning via out-of-distribution replay
G Kim, B Liu, Z Ke
Conference on Lifelong Learning Agents, 548-563, 2022
432022
Continual training of language models for few-shot learning
Z Ke, H Lin, Y Shao, H Xu, L Shu, B Liu
Proceedings of the 2022 Conference on Empirical Methods in Natural Language …, 2022
412022
Give me more feedback II: Annotating thesis strength and related attributes in student essays
Z Ke, H Inamdar, H Lin, V Ng
Proceedings of the 57th annual meeting of the association for computational …, 2019
392019
Parameter-level soft-masking for continual learning
T Konishi, M Kurokawa, C Ono, Z Ke, G Kim, B Liu
International Conference on Machine Learning, 17492-17505, 2023
362023
Continual learning with knowledge transfer for sentiment classification
Z Ke, B Liu, H Wang, L Shu
Machine Learning and Knowledge Discovery in Databases: European Conference …, 2021
362021
Bridging the Preference Gap between Retrievers and LLMs
Z Ke, W Kong, C Li, M Zhang, Q Mei, M Bendersky
Proceedings of the Annual Meeting of the Association for Computational …, 2024
252024
Adapting a language model while preserving its general knowledge
Z Ke, Y Shao, H Lin, H Xu, L Shu, B Liu
Proceedings of The 2022 Conference on Empirical Methods in Natural Language …, 2022
232022
Open-world continual learning: Unifying novelty detection and continual learning
G Kim, C Xiao, T Konishi, Z Ke, B Liu
Artificial Intelligence 338, 104237, 2025
172025
Sub-network Discovery and Soft-masking for Continual Learning of Mixed Tasks
Z Ke, B Liu, W Xiong, A Celikyilmaz, H Li
Findings of the Association for Computational Linguistics: EMNLP, 2023
72023
FaithEval: Can Your Language Model Stay Faithful to Context, Even If" The Moon is Made of Marshmallows"
Y Ming, S Purushwalkam, S Pandit, Z Ke, XP Nguyen, C Xiong, S Joty
arXiv preprint arXiv:2410.03727, 2024
62024
В данный момент система не может выполнить эту операцию. Повторите попытку позднее.
Статьи 1–20