AI Engineer
Cerebri AI
Toronto, ON, Canada
12h ago

Cerebri AI develops and sells the most sophisticated AI-enabled customer experience (CX) platform. Cerebri Values CX v2 (CVX) is a leap forward in applying artificial intelligence to measure customer engagement and drive financial results for large enterprises at scale.

It uses reinforcement learning to gain deeper insights from customer journeys, measure customer commitment across brands and predict Next Best Action set s to create personalized experiences throughout the customer life cycle that drive engagement and results.

We work with companies selling to over 200 million consumers and have 16 patents filed pertaining to the Cerebri AI way.

Our Series A financing was led by M12 (formerly Microsoft Ventures). We now have 60 employees in three offices in Austin, Toronto and Washington DC.

Over 80% of the staff are in technical roles in data science and software engineering. Our team of senior executives averages 20+ years selling and deploying software successfully to enterprises worldwide.

Cerebri AI is a proud Microsoft Partner and an active member of the Mastercard Start Path network.To learn more, visit cerebriai.

com. "Cerebri AI was recognized as 2019 Cool Vendor for Customer Journey Analytics by Gartner"

Responsibilities

  • Architect, build, test, deploy distributed, scalable, and resilient Spark / Scala / Kafka Big Data processing, and Machine Learning model pipelines for batch, micro-batch, and streaming workloads sets into Cerebri AI’s proprietary data stores for use in machine learning modeling
  • Develop and maintain data ontologies for key market segments
  • Collaborate with data scientists to develop automated orchestration of model pipelines to solve Cerebri AI business use case objectives
  • Collaborate with clients to develop pipeline infrastructure, and to ask appropriate questions to gain deep understanding of client data
  • Deploy fully containerized Docker / Kubernetes Data processing, and Machine Learning model pipelines into Azure, AWS, GCP cloud environments and on-premise systems as necessary
  • Document Detailed Designs (including source to target mappings) and Code for Data Quality frameworks that can measure and maintain Data Completeness, Data Integrity and Data Validity between interfacing systems
  • Ensure all solutions comply with the highest levels of security, privacy, and data governance requirements as outlined by Cerebri and Client legal and information security guidelines, law enforcement, and privacy legislation, including data anonymization, encryption, and security in transit and at rest, etc.
  • Train and mentor junior team members
  • Acts as a Subject Matter Expert and a Thought Leader, continuously following industry trends, the latest competitive developments, and delivering papers and presentations at major industry conferences and events.
  • Qualifications

  • A degree in Computer Science, Engineering, AI, Machine Learning, BI, MIS, or an equivalent technology field
  • Minimum 2 years of Production programming experience in Scala, Spark, PySpark, Big Data, Python
  • Minimum 2 years of Production experience with Hadoop Big Data platform
  • Able to program and understand data science and data engineering ideas in Python and translate into modular, functional components in Scala
  • Streaming and micro-batch application development experience would be an asset, including Kafka, Storm, NiFi, Spark Streaming, Confluent or equivalent
  • Proficiency with Linux / Unix operating systems, utilities and tools
  • Experience working directly with relational database structures and flat files
  • Ability to write efficient database queries, functions and views to include complex joins and the identification and development of custom indices
  • Knowledge of professional software engineering practices and best practices for the full software development life cycle, including coding standards, code reviews, source control management, build processes, testing, continuous integration and development, and operations.
  • Experience deploying containerized Docker / Kubernetes applications
  • Experience with Microsoft Azure or similar cloud computing solutions
  • Big Data application architecture experience and in-depth understanding of the Big Data ecosystem, applications, services, and design patterns
  • Production systems integration experience
  • Good verbal and written communication skills, with both technical and non-technical stakeholders
  • Nice to Haves

  • Experience in business intelligence visualization tools such as Grafana, Superset, Redash or Tableau.
  • Master’s degree or higher in a relevant quantitative subject
  • Experience with the Atlassian suite (JIRA, Confluence, BitBucket).
  • Any other related experience with Big Data, artificial intelligence, natural language processing, machine learning and / or deep learning, predictive analytics
  • Familiar with automated machine learning (AutoML) concepts would be an asset
  • Experience with Breeze would be an asset
  • Apply
    My Email
    By clicking on "Continue", I give neuvoo consent to process my data and to send me email alerts, as detailed in neuvoo's Privacy Policy . I may withdraw my consent or unsubscribe at any time.
    Continue
    Application form