CG-VAK Software & Exports Ltd.

Data Architect (Dremio Lakehouse)

Posted: 23 hours ago

Job Description

Role & ResponsibilitiesYou will be responsible for architecting, implementing, and optimizing Dremio-based data lakehouse environments integrated with cloud storage, BI, and data engineering ecosystems. The role requires a strong balance of architecture design, data modeling, query optimization, and governance enablement in large-scale analytical environments.Design and implement Dremio lakehouse architecture on cloud (AWS/Azure/Snowflake/Databricks ecosystem).Define data ingestion, curation, and semantic modeling strategies to support analytics and AI workloads.Optimize Dremio reflections, caching, and query performance for diverse data consumption patterns.Collaborate with data engineering teams to integrate data sources via APIs, JDBC, Delta/Parquet, and object storage layers (S3/ADLS).Establish best practices for data security, lineage, and access control aligned with enterprise governance policies.Support self-service analytics by enabling governed data products and semantic layers.Develop reusable design patterns, documentation, and standards for Dremio deployment, monitoring, and scaling.Work closely with BI and data science teams to ensure fast, reliable, and well-modeled access to enterprise data.Ideal CandidateBachelor’s or Master’s in Computer Science, Information Systems, or related field.5+ years in data architecture and engineering, with 3+ years in Dremio or modern lakehouse platforms.Strong expertise in SQL optimization, data modeling, and performance tuning within Dremio or similar query engines (Presto, Trino, Athena).Hands-on experience with cloud storage (S3, ADLS, GCS), Parquet/Delta/Iceberg formats, and distributed query planning.Knowledge of data integration tools and pipelines (Airflow, DBT, Kafka, Spark, etc.).Familiarity with enterprise data governance, metadata management, and role-based access control (RBAC).Excellent problem-solving, documentation, and stakeholder communication skills.PreferredExperience integrating Dremio with BI tools (Tableau, Power BI, Looker) and data catalogs (Collibra, Alation, Purview).Exposure to Snowflake, Databricks, or BigQuery environments.Experience in high-tech, manufacturing, or enterprise data modernization programs.Skills: dremio,modeling,enterprise,data,access control,design,cloud,storage,enterprise data,architecture

Job Application Tips

  • Tailor your resume to highlight relevant experience for this position
  • Write a compelling cover letter that addresses the specific requirements
  • Research the company culture and values before applying
  • Prepare examples of your work that demonstrate your skills
  • Follow up on your application after a reasonable time period

You May Also Be Interested In