Toward building more accessible large language models: A preliminary empirical study on data scarcity in knowledge distillation and algorithm complexity in alignment