Senior Software Engineer - Data Platform
Join our stealth initiative to architect next-generation data unification platforms that transform how enterprises leverage their data assets through digital twin technology and semantic ontologies.
Key Responsibilities
- Design and implement high-performance data ingestion pipelines processing structured and unstructured data from heterogeneous sources
- Develop digital twin architectures that create real-time virtual representations of physical and logical data entities
- Build semantic ontology engines that automatically discover, classify, and establish relationships between disparate data elements
- Implement graph-based data models and knowledge graphs to represent complex entity relationships and hierarchies
- Create data virtualization layers enabling unified access to distributed data without physical consolidation
- Develop API frameworks and SDKs that enable rapid development of data-driven applications
- Optimize query engines for sub-second response times across petabyte-scale federated datasets
- Build ML pipelines for automated metadata extraction, entity resolution, and relationship inference
- Implement data lineage tracking and impact analysis capabilities across the unified data fabric
- Design security and governance frameworks ensuring data privacy while maximizing accessibility
Requirements
- Bachelor's degree in Computer Science, Software Engineering, or related technical field
- 7+ years of experience building distributed systems and data platforms at scale
- Expert-level proficiency in multiple programming languages including Golang, C/C++, Rust, and Python
- Deep understanding of distributed computing principles, CAP theorem, and consensus algorithms
- Experience with graph databases (Neo4j, Amazon Neptune, ArangoDB) and graph processing frameworks
- Strong background in semantic web technologies (RDF, OWL, SPARQL) and knowledge representation
- Proven track record building data integration platforms handling diverse data formats and protocols
- Experience with stream processing frameworks (Apache Kafka, Pulsar, Flink) and event-driven architectures
- Expertise in database internals, query optimization, and distributed query execution
- Understanding of machine learning fundamentals and experience with ML frameworks (PyTorch, TensorFlow)
Preferred Qualifications
- Advanced degree in Computer Science, Data Science, or related field
- Experience with digital twin implementations in industrial IoT or smart city contexts
- Knowledge of industry-specific data standards (FHIR for healthcare, FIBO for finance, CIM for utilities)
- Contributions to open-source data platform or database projects
- Experience with vector databases and embedding-based similarity search
- Background in compiler design, query planning, or database optimizer development
- Familiarity with data mesh architectures and federated governance models
- Published research in data integration, knowledge graphs, or semantic technologies
Benefits & Perks
- Competitive base salary ranging from $180,000 to $250,000 based on experience
- Performance-based bonus up to 35% of base salary
- Equity participation in groundbreaking technology initiatives
- Comprehensive health, dental, and vision insurance
- 401(k) with 6% company match
- $10,000 annual budget for conferences, training, and professional development
- Access to cutting-edge hardware and cloud resources
- Flexible PTO and sabbatical options
- Patent filing support and innovation bonuses
About This Role
Jane Isis is seeking a Senior Software Engineer to join Project Labradoor, our ambitious initiative to revolutionize enterprise data management. This stealth project aims to solve one of the most persistent challenges in modern organizations: the fragmentation of valuable data across siloed systems, formats, and domains. You’ll be at the forefront of developing technology that creates a unified semantic layer across all enterprise data, enabling unprecedented insights and capabilities.
The Vision: Breaking Down Data Silos
Modern enterprises struggle with data trapped in isolated systems—ERP databases that don’t communicate with IoT sensors, unstructured documents disconnected from transactional records, and machine-generated logs isolated from business analytics. Project Labradoor addresses this challenge by creating a revolutionary data unification platform that:
- Digital Twin Creation: Generates virtual representations of all data entities, whether they originate from physical sensors, business applications, or unstructured content
- Semantic Ontology Development: Automatically constructs domain-specific knowledge graphs that capture the relationships, hierarchies, and business rules governing data entities
- Intelligent Data Linking: Employs advanced entity resolution and relationship inference to connect previously isolated data points across systems
- Unified Access Layer: Provides a single, coherent interface to query and interact with all enterprise data, regardless of its original source or format
Technical Challenges You’ll Solve
This role demands solving complex distributed systems challenges at the intersection of data engineering, knowledge representation, and high-performance computing:
System Architecture: Design distributed systems capable of ingesting millions of events per second while maintaining ACID guarantees for critical operations. Build polyglot persistence layers that optimize storage and retrieval for different data characteristics—time-series data in columnar stores, relationships in graph databases, and documents in search indices.
Performance Engineering: Implement lock-free data structures and zero-copy techniques in C++ and Rust for ultra-low latency data processing. Optimize memory layouts for cache efficiency and SIMD vectorization. Build custom allocators and memory pools to minimize garbage collection overhead in high-throughput scenarios.
Semantic Intelligence: Develop ontology inference engines that automatically discover conceptual models from raw data. Implement description logic reasoners for automated classification and consistency checking. Build SPARQL query processors that execute complex semantic queries across distributed knowledge graphs.
Language Diversity: Leverage each language’s strengths—Golang for concurrent microservices and API development, Rust for memory-safe systems programming, C++ for performance-critical components, and Python for ML pipelines and rapid prototyping.
Your Impact
Your work will fundamentally transform how organizations leverage their data assets:
- Accelerate AI/ML Adoption: Enable data scientists to access all relevant data through a unified semantic layer, reducing feature engineering time from months to days
- Enable Real-time Intelligence: Power applications that combine streaming IoT data with historical records and unstructured content for immediate insights
- Reduce Integration Costs: Eliminate the need for point-to-point integrations by providing a universal data access layer
- Unlock Hidden Value: Discover previously unknown relationships and patterns by connecting isolated data silos
- Democratize Data Access: Enable business users to query complex data relationships using intuitive semantic models
The Technology Stack
You’ll work with cutting-edge technologies across the entire data platform spectrum:
- Languages: Golang (microservices, APIs), Rust (core engine, performance-critical paths), C++ (query optimization, memory management), Python (ML pipelines, orchestration)
- Data Stores: Graph databases, time-series databases, vector databases, object stores, streaming platforms
- Frameworks: Apache Arrow for columnar processing, gRPC for service communication, WebAssembly for portable compute
- Infrastructure: Kubernetes for orchestration, service mesh for microservice communication, observability platforms for monitoring
Team & Culture
You’ll join a small, elite team of engineers who’ve built systems at companies like Google, Amazon, and Palantir. Our team includes experts in distributed systems, database internals, semantic web technologies, and machine learning. We value deep technical expertise, intellectual curiosity, and the ability to tackle ambiguous problems with creative solutions.
Growth Opportunities
This role offers unique career growth potential. As an early member of Project Labradoor, you’ll have the opportunity to shape the technical direction of a platform that could redefine enterprise data management. Successful delivery could lead to technical leadership roles, patent authorship, and the chance to present at major conferences as we bring this technology to market.
Why This Role Matters
Data silos cost enterprises billions in missed opportunities and inefficiencies. By joining Project Labradoor, you’ll help build the foundational technology that enables organizations to finally realize the full value of their data assets. This isn’t just another data platform—it’s a fundamental reimagining of how enterprises understand and interact with their information.
Application Process
Ready to break down data silos and build the future of enterprise data platforms? Submit your resume, GitHub profile, and a brief description of the most challenging distributed systems problem you’ve solved. We’re particularly interested in examples demonstrating your expertise across multiple programming languages and your ability to design systems that balance performance, scalability, and maintainability.
Apply for This Position
Take the next step in your career. Submit your application and we'll review it within 48 hours.