The research focus is to apply the pre-training techniques of Large Language Models to the encoding process of the Code Search Project, to improve the existing model and develop a new code searching model. The assistant shall explore a transformer or equivalent model (such as GPT-3.5) with fine-tuning, which can help achieve state-of-the-art performance for NLP tasks. The research also aims to test and evaluate various state-of-the-art models to find the most promising ones.
I am not familiar with exchanging credits for hours on machine and any guidance as which cluster and how many hours needed on it would be great. I am experienced in python coding and would like to work on a python friendly interface and i need to parallelize my code to run for lots of different parameter values and it also involves large matrix so quite some memory overhead.