Deprecated: Constant E_STRICT is deprecated in /var/www/html/system/core/Exceptions.php on line 75

A PHP Error was encountered

Severity: 8192

Message: Creation of dynamic property CI_Cache::$apc is deprecated

Filename: libraries/Driver.php

Line Number: 188

Backtrace:

File: /var/www/html/application/helpers/global_helper.php
Line: 561
Function: driver

File: /var/www/html/application/helpers/global_helper.php
Line: 594
Function: getCategories

File: /var/www/html/application/controllers/Jobs.php
Line: 290
Function: getCategoryDetByCatId

File: /var/www/html/index.php
Line: 315
Function: require_once

A PHP Error was encountered

Severity: 8192

Message: Creation of dynamic property CI_Cache::$file is deprecated

Filename: libraries/Driver.php

Line Number: 188

Backtrace:

File: /var/www/html/application/helpers/global_helper.php
Line: 561
Function: driver

File: /var/www/html/application/helpers/global_helper.php
Line: 594
Function: getCategories

File: /var/www/html/application/controllers/Jobs.php
Line: 290
Function: getCategoryDetByCatId

File: /var/www/html/index.php
Line: 315
Function: require_once

A PHP Error was encountered

Severity: 8192

Message: Creation of dynamic property CI_Cache::$dummy is deprecated

Filename: libraries/Driver.php

Line Number: 188

Backtrace:

File: /var/www/html/application/helpers/global_helper.php
Line: 562
Function: get

File: /var/www/html/application/helpers/global_helper.php
Line: 594
Function: getCategories

File: /var/www/html/application/controllers/Jobs.php
Line: 290
Function: getCategoryDetByCatId

File: /var/www/html/index.php
Line: 315
Function: require_once

Azure Databricks Engineer
  • VRK IT VISION
76 Days Ago
55-58 per Hourly
NA
Remote
10-18 Years
Required Skills: Databricks, Pyspark, Azure Cloud Services, Asset Bundle
Job Description
Role : Azure Databricks Engineer
Location : New York, NY Onsite – day 1
  
Must have : Databricks , Pyspark , Azure Cloud Services, Asset Bundle
 
Technical Skills:
  • Strong expertise in Databricks (Delta Lake, Unity Catalog, Lakehouse Architecture, Table Triggers, Delta Live Pipelines, Databricks Runtime etc.)
  • Proficiency in Azure Cloud Services.
  • Solid Understanding of Spark and PySpark for big data processing.
  • Experience in relational databases.
  • Knowledge on Databricks Asset Bundles and GitLab.
 
Key Responsibilities:
  1. Data Pipeline Development:
    • Build and maintain scalable ETL/ELT pipelines using Databricks.
    • Leverage PySpark/Spark and SQL to transform and process large datasets.
    • Integrate data from multiple sources including Azure Blob Storage, ADLS and other relational/non-relational systems.
  2. Collaboration & Analysis:
    • Work Closely with multiple teams to prepare data for dashboard and BI Tools.
    • Collaborate with cross-functional teams to understand business requirements and deliver tailored data solutions.
  3. Performance & Optimization:
    • Optimize Databricks workloads for cost efficiency and performance.
    • Monitor and troubleshoot data pipelines to ensure reliability and accuracy.
  4. Governance & Security:
    • Implement and manage data security, access controls and governance standards using Unity Catalog.
    • Ensure compliance with organizational and regulatory data policies.
  5. Deployment:
    • Leverage Databricks Asset Bundles for seamless deployment of Databricks jobs, notebooks and configurations across environments.
    • Manage version control for Databricks artifacts and collaborate with team to maintain development best practices.
 
Preferred Experience:
  • Familiarity with Databricks Runtimes and advanced configurations.
  • Knowledge of streaming frameworks like Spark Streaming.
  • Experience in developing real-time data solutions.
 
Certifications:
  • Azure Data Engineer Associate or Databricks certified Data Engineer Associate certification. (Optional)

Jobseeker

Looking For Job?
Search Jobs

Recruiter

Are You Recruiting?
Search Candidates
TekJobs All Rights Reserved

https://www.tekjobs.net/jobs/applyjob", type: "POST", enctype: 'multipart/form-data', data: formData, processData: false, contentType: false, cache: false, // timeout: 60000, success: function(data) { const jsonData = JSON.parse(data); if(jsonData.hasOwnProperty('error')) { alert(jsonData.error); } else if(jsonData.hasOwnProperty('success')) { alert('successfully applied for this job'); $(".close").trigger('click'); $("#name").val(''); $("#contactNumber").val(''); $("#resume").val(''); $("#fromEmail").trigger('focusout'); } else { setTimeout(()=> { alert('something went wrong'); // window.location.reload(); }, 2000); } } }) return false; } }); $("#btnSendMail").hide(); $("#selectedResume").on('change', function() { if($("#selectedResume").val().length) { $("#btnSendMail").show(); } else { $("#btnSendMail").hide(); } }); $("#fromEmail").on('focusout', function() { $("#btnSendMail").hide(); appliedEmail = $("#fromEmail").val(); if(appliedEmail.trim() == '') { $("#resumeSelect").hide(); $("#fillInfo1").show(); $("#fillInfo2").show(); $("#p-fill").hide(); } mailFormat = /\S+@\S+\.\S+/; if(mailFormat.test(appliedEmail)) { // alert('hello'); $.ajax({ url: "https://www.tekjobs.net/candidate/getCandidatesDetail", data: {email: appliedEmail}, type: "POST", success: function(data) { jsonData = JSON.parse(data); if(jsonData.hasOwnProperty('candDetail')) { if(jsonData.candDetail.length == 0) { $("#selectedResume").html(''); $("#resumeSelect").hide(); $("#fillInfo1").show(); $("#fillInfo2").show(); $("#p-fill").hide(); } else { let htmlCont = ''; $.each(jsonData.candDetail, function(key, value) { htmlCont += ''; }); $("#selectedResume").html(htmlCont); $("#resumeSelect").show(); $("#fillInfo1").show(); $("#fillInfo2").hide(); $("#p-fill").show(); } } else { $("#selectedResume").html(''); } $("#selectedResume").multiselect('rebuild'); $("#formApplyJob").valid(); } }); } }) $("#selectedResume,#name,#contactNumber,#resume").on('change', ()=> { $("#formApplyJob").valid(); }) const emailCookie = $("#fromEmail").val(); if(emailCookie) { $("#fromEmail").trigger('focusout'); } });