Job
Description
As an AIML - LLM Gen AI Engineer, your primary responsibility will be to design, develop, and implement solutions using transformer-based models for various Natural Language Processing (NLP) tasks. This includes tasks such as text generation, summarization, question answering, classification, and translation. You will be working extensively with transformer models like GPT, BERT, T5, RoBERTa, and similar architectures. Your deep understanding of model architectures, attention mechanisms, and self-attention layers will be crucial in effectively utilizing Large Language Models (LLMs) to generate human-like text. You will lead efforts in fine-tuning pre-trained LLMs and other transformer models on domain-specific datasets to optimize performance for specialized NLP tasks. In your role, you will apply your knowledge of attention mechanisms, context windows, tokenization, and embedding layers in model development and optimization. It will be essential to address and mitigate issues related to biases, hallucinations, and knowledge cutoffs that can impact LLM performance and output quality. You will craft clear, concise, and contextually relevant prompts to guide LLMs towards generating desired outputs, including the use of instruction-based prompting. Additionally, you will implement and experiment with zero-shot, few-shot, and many-shot learning techniques to maximize model performance without extensive retraining. Your iterative approach to prompt engineering strategies will involve refining outputs, rigorously testing model performance, and ensuring consistent and high-quality results. You will also be responsible for creating prompt templates for repetitive tasks that are adaptable to different contexts and inputs. Expertise in chain-of-thought (CoT) prompting will enable you to guide LLMs through complex reasoning tasks by encouraging step-by-step breakdowns. Your contribution will span the entire lifecycle of machine learning models in an NLP context, including training, fine-tuning, and deployment. **Required Skills & Qualifications:** - A minimum of 8 years of experience working with transformer-based models and NLP tasks, focusing on text generation, summarization, question answering, classification, and similar applications. - Proficiency in transformer models such as GPT, BERT, T5, RoBERTa, and foundational models. - Strong familiarity with model architectures, attention mechanisms, and self-attention layers for generating human-like text. - Proven experience in fine-tuning pre-trained models on domain-specific datasets for various NLP tasks. - Knowledge of attention mechanisms, context windows, tokenization, and embedding layers. - Awareness of biases, hallucinations, and knowledge cutoffs affecting LLM performance. - Ability to craft clear, concise, and contextually relevant prompts for LLMs. - Experience in zero-shot, few-shot, and many-shot learning techniques. - Proficiency in Python and NLP libraries like Hugging Face Transformers, SpaCy, NLTK. - Solid experience in training, fine-tuning, and deploying ML models in an NLP context. - Strong problem-solving skills and analytical mindset. - Excellent communication and collaboration abilities for remote or hybrid work environments. - Bachelor's or Master's degree in Computer Science, Artificial Intelligence, Machine Learning, or related quantitative field. This role requires a combination of technical expertise in AI/ML, NLP, and transformer models, along with strong problem-solving and communication skills to achieve effective results in the field.,