EGeeks Global is seeking a Generative AI Engineer with 1.5–2 years of experience in building and optimizing advanced AI models, specifically within the realm of Large Language Models (LLMs), such as Meta’s LLaMA, GPT-4, or Falcon. The ideal candidate will have hands-on experience in GPU-based optimization, vector database integration, and AI-driven Robotic Process Automation (RPA) workflows. In this role, you’ll leverage cutting-edge AI tools and frameworks to tackle complex business challenges and contribute to AI-driven automation projects that align with future trends.
Key Responsibilities
As a Generative AI Engineer, your role will focus on several areas:
1. AI Model Development and Optimization
- Design and fine-tune LLaMA-based generative AI models for domain-specific applications.
- Implement optimization techniques (e.g., LoRA, quantization, PEFT) to customize and improve model performance.
- Leverage GPU-accelerated frameworks (e.g., CUDA, TensorRT, PyTorch) for training and optimizing models.
2. NLP and LLM Applications
- Adapt and implement LLMs (like GPT-4, LLaMA, Gemini) for various applications, including text summarization, chatbot development, and semantic search.
- Develop NLP models for tasks such as intent recognition, sentiment analysis, and entity extraction.
3. Vector Database Integration
- Integrate vector databases (e.g., FAISS, Pinecone, Milvus, ChromaDB) for efficient embedding storage and retrieval.
- Develop embedding generation pipelines for real-time AI applications.
4. GPU-Based Performance Enhancement
- Utilize NVIDIA GPUs, TensorRT, and ONNX Runtime to optimize model deployments for low-latency, high-throughput scenarios.
- Enhance model scalability, efficiency, and cost-effectiveness in production environments.
5. RPA and AI-Driven Automation
- Develop intelligent systems combining LLMs and RPA platforms (e.g., UiPath, Automation Anywhere) to automate complex business workflows.
- Design AI agents to enhance task automation, improving overall efficiency.
6. Cloud Deployment and Integration
- Deploy AI models to cloud platforms (e.g., AWS SageMaker, Azure AI, Google AI Platform).
- Integrate AI models into APIs, microservices, and production pipelines.
7. Monitoring, Maintenance, and Optimization
- Monitor and optimize deployed models based on performance, accuracy, and user feedback.
- Continuously update knowledge on emerging AI tools and technologies, integrating them into workflows.
Required Skills and Qualifications
We are seeking candidates who meet the following qualifications and experience:
- Education: Bachelor’s degree in Computer Science, AI, Machine Learning, or related fields.
- Experience: 1.5 to 2 years of practical experience in AI/ML, especially with LLMs, NLP, and deploying AI models.
- Technical Proficiency:
- Strong experience with frameworks like PyTorch, TensorFlow, and Hugging Face Transformers.
- Practical knowledge of models like Meta’s LLaMA, GPT-4, or Falcon.
- Hands-on experience with vector databases (e.g., FAISS, Pinecone, Milvus).
- Proficient with GPU-based development tools such as CUDA, TensorRT, ONNX Runtime.
- Familiarity with RPA platforms (e.g., UiPath, Automation Anywhere) and AI-driven automation techniques.
- Deployment Skills: Experience with cloud deployment (e.g., AWS, Azure, GCP).
Preferred Skills
While not required, the following skills will give candidates a competitive advantage:
- Orchestration tools: Exposure to LangChain or Auto-GPT for managing AI agents.
- Distributed training frameworks like Ray or DeepSpeed.
- Familiarity with MLOps and lifecycle management for LLMs.
- Experience with reinforcement learning and agent-based systems.
- Experience with experiment tracking tools like Weights & Biases (WandB).
What We Offer
- Cutting-Edge Technology: Gain experience working with leading AI technologies such as LLaMA, GPT-4, vector databases, and AI-driven automation.
- Collaborative Work Environment: Join a team dedicated to advancing innovative AI solutions and explore real-world deployments across scalable cloud platforms.
- Career Development: Work on impactful, real-world AI projects and position yourself at the forefront of AI advancements for the years to come.
Application Instructions
If you’re ready to bring your AI expertise to the table and work on groundbreaking technologies, we encourage you to apply today:
- How to Apply: Submit your CV to hr@egeeksglobal.com. Make sure your CV highlights relevant work experience, technical skills, qualifications, and certifications.
- Location: Rawalpindi, Punjab, Pakistan (On-site role).
- Associate Product Manager (Scrum Master) – Join BoolMind in Lahore - January 23, 2025
- Sr. Factory Reporting & Operations Specialist – Haier Pakistan - January 23, 2025
- Senior Education Consultant – ABN Overseas Education - January 23, 2025