Easy Apply Now

A bit about us:

Excellent Private Equity Backed Software Organization developing AI-Driven Solutions for Critical Infrastructure

100% REMOTE

Must Have Vector Database Experience

Why join us?

100% Remote

Meaningful Work

Creativity

Excellent Benefits

Job Details

Full Time - Remote

Base: $170-200k + Bonus


Job Details:

We are seeking a dynamic and innovative Data Architect to join our GenAI team. This is a full-time, permanent position that can be done remotely. The successful candidate will be part of a fast-paced, entrepreneurial team responsible for accelerating and driving the growth of our cutting-edge AI-driven technology services. With a focus on data architecture, the role involves designing, creating, deploying, and managing our organization's data architecture. This includes ETL processes, SQL and NoSQL databases, data streaming, and real-time data pipelines.

Responsibilities:

  • Design, create, and manage the data architecture, including data models, database design, data integration, storage, and data partitioning.
  • Develop and implement scalable ETL processes, ensuring the accuracy and integrity of data.
  • Work with SQL, NoSQL, and distributed databases to ensure optimal performance and data availability.
  • Utilize data streaming technologies such as Flink and Kinesis to manage real-time data pipelines.
  • Implement AI-driven technologies like Pinecone, Milvus, Weaviate, and OpenAI GPT into the data architecture.
  • Create and manage data partitioning strategies to optimize database performance.
  • Use Python to automate processes and integrate systems.
  • Work with Vector and other SaaS tools to manage data.
  • Collaborate with the AI team to implement and manage AI-driven solutions.
  • Develop and manage data embeddings to optimize data retrieval and usage.

Qualifications:

  • Bachelor's degree in Computer Science, Information Technology, or a related field.
  • A minimum of 5 years of experience as a Data Architect or in a similar role.
  • Extensive experience with ETL, SQL, NoSQL, data streaming, and real-time data pipelines.
  • Proficiency in Python is a must.
  • Experience with Kafka, Flink, Kinesis, etc.
  • Knowledge of Pinecone, Milvus, Weaviate, OpenAI GPT, and other AI-driven technologies is highly desirable.
  • Familiarity with data partitioning and distributed databases is necessary.
  • Strong knowledge of data embeddings is a plus.
  • Excellent problem-solving skills and attention to detail.
  • Strong communication and collaboration skills.
  • Ability to work independently and manage multiple tasks simultaneously.
  • Proven ability to work in a remote environment.
Easy Apply Now