Position Overview<\/b>
<\/div><\/p>
We are looking for a highly skilled Senior Generative AI
/ Backend Engineer<\/b> with strong experience in building AI -powered
applications and scalable backend systems. The ideal candidate will have
hands -on expertise in LLMs, Retrieval -Augmented Generation (RAG), prompt
engineering<\/b>, and Python -based backend development<\/b>, with the ability
to deliver production -ready AI solutions.
<\/p>
Key Responsibilities<\/b>
<\/p>
Generative AI Development<\/b>
<\/p>- Design,
develop, and deploy generative AI solutions using frameworks such as LangChain<\/b> and LlamaIndex<\/b>.
<\/li>- Apply
advanced prompt engineering<\/b> techniques to optimize LLM responses
across multiple use cases.
<\/li>- Work
with advanced LLM features including prompt optimization,
hyperparameter tuning, and response caching<\/b>.
<\/li>- Implement Retrieval -Augmented Generation (RAG)<\/b> workflows by integrating
vector databases such as Pinecone, Weaviate, Supabase, or PGVector<\/b>.
<\/li>- Develop
solutions leveraging embeddings and similarity search<\/b> for
intelligent and personalized query resolution.
<\/li>- Process
and analyze multimodal data<\/b>, including text, images, and video for
AI -driven applications.
<\/li>- Integrate observability and monitoring tools<\/b> to track LLM performance,
reliability, and quality.
<\/li><\/ul>Backend Engineering<\/b>
<\/p>- Build
and maintain scalable, secure backend services using Python frameworks<\/b> such as FastAPI, Django, or Flask<\/b>.
<\/li>- Design
and implement RESTful APIs<\/b> for seamless system integration.
<\/li>- Optimize
and manage relational databases<\/b> (PostgreSQL, MySQL) and vector
databases<\/b> for AI workflows.
<\/li>- Implement asynchronous programming<\/b> patterns and follow clean coding best
practices.
<\/li>- Integrate third -party SDKs and APIs<\/b> to support external system
interoperability.
<\/li>- Develop
backend pipelines for multimodal data processing<\/b> (text, image, and
video).
<\/li>- Manage
background and scheduled tasks using Celery, cron jobs, or job queues<\/b>.
<\/li>- Use Docker<\/b> for containerization and reproducible deployments.
<\/li>- Ensure
backend systems are secure, scalable, and production -ready<\/b>.
<\/li><\/ul>Qualifications<\/b>
<\/p>
Essential Skills<\/b>
<\/p>- Strong
proficiency in Python<\/b> and backend frameworks (FastAPI, Django,
Flask<\/b>).
<\/li>- Hands -on
experience with Generative AI frameworks<\/b> such as LangChain,
LlamaIndex<\/b>, and RAG architectures.
<\/li>- Strong
understanding of LLMs, embeddings, and similarity search<\/b> techniques.
<\/li>- Experience
with relational databases<\/b> (PostgreSQL, MySQL) and vector
databases<\/b> (Pinecone, Weaviate, Supabase, PGVector).
<\/li>- Experience
deploying AI solutions to production using Docker<\/b>.
<\/li>- Strong
skills in testing, debugging, and performance optimization<\/b> (Pytest,
unit/integration testing).
<\/li>- Solid
understanding of asynchronous programming<\/b> and concurrent task
handling.
<\/li><\/ul>Preferred Skills<\/b>
<\/p>- Experience
with cloud platforms<\/b> such as AWS, GCP, or Azure<\/b>.
<\/li>- Basic
understanding of frontend technologies<\/b> (HTML, CSS, JavaScript;
Angular or React is a plus).
<\/li>- Experience
with observability and monitoring tools<\/b> for real -time LLM
evaluation.
<\/li>- Awareness
of emerging trends in generative AI, multimodal AI, and agent -based
workflows<\/b>.
<\/li>- Knowledge
of secure coding practices<\/b> and backend system hardening.
<\/li>- Relevant certifications<\/b> in AI, Machine Learning, or Cloud technologies are a
plus.
<\/li><\/ul>
<\/div><\/span>