Senior Software Engineer – Foundational Data Systems for AI - Canada

This listing is synced directly from the company ATS.

Role Overview

This senior-level role involves designing and building foundational data systems for AI, including global metadata substrates, adaptive engines, and intelligent data layouts. The engineer will work on distributed systems at petabyte scale, implementing algorithms from research into production to optimize data efficiency and support enterprise AI workloads. They will have a significant impact by shaping the core infrastructure that makes AI more efficient and sustainable.

Perks & Benefits

The job offers remote work with flexible time off, comprehensive health coverage for the family, and support for research and publication. It provides a high-trust, low-bureaucracy environment with real ownership over long-term projects, competitive salary, meaningful equity, and bonuses for top performers. Career growth includes opportunities to work at the intersection of science and engineering, with backing from major investors for enduring impact.

⚠️ This job was posted over 3 months ago and may no longer be open. We recommend checking the company's site for the latest status.

Full Job Description

About Granica

Granica is an AI research and infrastructure company focused on reliable, steerable representations for enterprise data.

We earn trust through Crunch, a policy-driven health layer that keeps large tabular datasets efficient, reliable, and reversible. On this foundation, we’re building Large Tabular Models—systems that learn cross-column and relational structure to deliver trustworthy answers and automation with built-in provenance and governance.

The Mission

AI today is limited not only by model design but by the inefficiency of the data that feeds it. At scale, each redundant byte, each poorly organized dataset, and each inefficient data path slows progress and compounds into enormous cost, latency, and energy waste.

Granica’s mission is to remove that inefficiency. We combine new research in information theory, probabilistic modeling, and distributed systems to design self-optimizing data infrastructure: systems that continuously improve how information is represented and used by AI.

This engineering team partners closely with the Granica Research group led by Prof. Andrea Montanari (Stanford), bridging advances in information theory and learning efficiency with large-scale distributed systems. Together, we share a conviction that the next leap in AI will come from breakthroughs in efficient systems, not just larger models.

What You’ll Build

  • Global Metadata Substrate. Architect the transactional and metadata substrate that supports time-travel, schema evolution, and atomic consistency across petabyte-scale tabular datasets.

  • Adaptive Engines. Build systems that reorganize data autonomously, learning from access patterns and workloads to maintain peak efficiency without manual tuning.

  • Intelligent Data Layouts. Optimize bit-level organization (encoding, compression, layout) to extract maximal signal per byte read.

  • Autonomous Compute Pipelines. Develop distributed compute systems that scale predictively, adapt to dynamic load, and maintain reliability under failure.

  • Research to Production. Implement new algorithms in compression, representation, and optimization emerging from ongoing research. Opportunities to publish and open-source are encouraged.

  • Latency as Intelligence. Design for minimal time between question and insight, enabling models and humans to learn faster from data.

What You Bring

  • Depth in distributed systems: consensus, partitioning, replication, fault tolerance.

  • Experience with columnar formats such as Parquet or ORC and low-level encoding strategies.

  • Understanding of metadata-driven architectures and adaptive query planning.

  • Production experience with Spark, Flink, or custom distributed engines on cloud object storage.

  • Proficiency in Java, Rust, Go, or C++ with an emphasis on clarity and quality.

  • Curiosity about theory of the mathematics of compression, entropy, and learning efficiency.

  • A builder’s mindset: pragmatic, rigorous, and grounded in long-term systems thinking.

Bonus

  • Familiarity with Iceberg, Delta Lake, or Hudi.

  • Research or open-source contributions in compression, indexing, or distributed computation.

  • Interest in how data representation affects training dynamics and model reasoning efficiency.

Why Granica

  • Fundamental Research Meets Enterprise Impact. Work at the intersection of science and engineering, turning foundational research into deployed systems serving enterprise workloads at exabyte scale.

  • AI by Design. Build the infrastructure that defines how efficiently the world can create and apply intelligence.

  • Real Ownership. Design primitives that will underpin the next decade of AI infrastructure.

  • High-Trust Environment. Deep technical work, minimal bureaucracy, shared mission.

  • Enduring Horizon. Backed by NEA, Bain Capital, and various luminaries from tech and business. We are building a generational company for decades, not quarters or a product cycle.

Compensation & Benefits

  • Competitive salary, meaningful equity, and substantial bonus for top performers

  • Flexible time off plus comprehensive health coverage for you and your family

  • Support for research, publication, and deep technical exploration

At Granica, you will shape the fundamental infrastructure that makes intelligence itself efficient, structured, and enduring. Join us to build the foundational data systems that power the future of enterprise AI!

Similar jobs

Found 6 similar jobs