Titre du poste ou emplacement
RECHERCHES RÉCENTES

Sr. Data Architect with GCP-Canada

Zortech Solutions - 17 emplois
Toronto, ON
Publié il y a 7 jours
Détails de l'emploi :
Temps plein
Expérimenté

Role: Sr. Data Architect with GCPLocation: Remote-CanadaDuration: 6-12+ MonthsJob Description :Key Responsibilities:
  • Join our quest at client as a Data Architect, where you'll be at the heart of innovation, shaping our trailblazing Generative AI platform. Our vision? To revolutionize the Gen-AI landscape, making it as universal and user-friendly as possible, touching lives from seasoned developers to creative designers and everyone in between.
  • Our mantra is simple yet profound: 'Focus, Flow, and Joy'. If you have a fervent interest in crafting innovative products for a broad audience and are excited about leveraging state-of-the-art technology, this is the right role for you.
  • Imagine being part of our Data Platform Engineering team, a place where status quo are questioned, processes are perfected, and cutting-edge tech is used not just for the sake of it, but to fundamentally transform effort into efficiency, and ideas into reality. This isn't just a job; it's a journey to redefine the future of technology.
What you'll do
  • You will be part of a dynamic team with experienced professionals, dedicated to delivering comprehensive solutions that harness the power of Data and Generative AI technology, including the development of custom-built products.
  • By joining our community of passionate builders, you will contribute to our shared goal of providing the most valuable, user-friendly, and enjoyable experiences! You will be playing a key role in ensuring the quality and rapid delivery of the products built using the client Generative AI platform.
We enjoy:
  • Exploring bleeding-edge technologies, tools and frameworks to experiment with and build better products for existing customers
  • Evaluating areas of improvement with technical products built and implementing ideas which will make us better than yesterday
  • Collaborating with developers to work on technical designs and develop code, configurations, and scripts to enhance the development lifecycle and integrate systems
  • Collaborate proactively and respectfully with our team and customers
  • Develop tools and integrations to support other developers in building products
  • Take solutions from concept to production by writing code, configurations, and scripts
  • Improve existing platforms or implement new features for any of our products
  • Create comprehensive documentation for implemented solutions, including implementation details and usage instructions
  • Promote our culture of focus, flow, and joy to gain developers' support for our solutions
Qualifications
  • What you bring
  • Build Data pipelines required for optimal extraction, anonymization, and transformation of data from a wide variety of data sources using SQL, NoSQL and AWS 'big data' technologies.
    • Streaming
    • Batch
  • Work with stakeholders including the Product Owners, Developers and Data scientists to assist
  • with data-related technical issues and support their data infrastructure needs.
  • Ensure that data is secure and separated following corporate compliance and data governance
  • policies
  • Take ownership of existing ETL scripts, maintain and rewrite them in modern data
  • transformation tools whenever needed.
  • Being an automation advocate for data transformation, cleaning and reporting tools.
  • You are proficient in developing software from idea to production
  • You can write automated test suites for your preferred language
  • You have frontend development experience with frameworks such as React.js/Angular
  • You have backend development experience building and integrating with REST APIs and Databases using languages such as Java Spring, JavaScript on Node.js, Flask on Python
  • You have experience with cloud-native technologies, such as Cloud Composer, Dataflow, Dataproc, BigQuery, GKE, Cloud run, Docker, Kubernetes, and Terraform
  • You have used cloud platforms such as Google Cloud/AWS for application hosting
  • You have used and understand CI/CD best practices with tools such as GitHub Actions, GCP Cloud Build
  • You have experience with YAML and JSON for configuration
  • You are up-to-date on the latest trends in AI Technology
Great-to-haves:
  • 3+ years of experience as a data or software architect
  • 3+ years of experience in SQL and Python
  • 2+ years of experience with ELT/ETL platforms (Airflow, DBT, Apache Beam, PySpark, Airbyte)
  • 2+ years of experience with BI reporting tools (Looker, Metabase, Quicksight, PowerBI, Tableau)
  • Extensive knowledge of the Google Cloud Platform, specifically the Google Kubernetes Engine
  • Experience with GCP cloud data related services ( Dataflow, GCS, Datastream, Data Fusion, Data Application, BigQuery, Data Flow, Data Proc, Dataplex, PubSub, CloudSQL, BigTable)
  • Experience in health industry an asset
  • Expertise in Python, Java
  • Interest in PaLM, LLM usage and LLMOps
  • Familiarity with LangFuse or Backstage plugins or GitHub Actions
  • Strong experience with GitHub beyond source control
  • Familiarity with monitoring, alerts, and logging solutions
  • Join us on this exciting journey to make Generative AI accessible to all and create a positive impact with technology

#L!-CEIPAL

Partager un emploi :