Background:
We are seeking a Subject Matter Expert (SME) in Big Data & Analytics to lead the design, development, and optimization of data-driven solutions. The ideal candidate will have deep experience in big data technologies, data pipelines, and advanced analytics to drive business intelligence, predictive modeling, and strategic decision-making.
Scope of Work:
● Create a course structure for a certificate program with 4-5 courses (number of courses to be based on scoping). Each course is likely to have 4-5 modules and a total of 25 lessons. So a 4-course program could have up to 100 lessons.
● Work closely with the client in a rigorous scoping process to create a Job Task Analysis document and content structure for each program and course.
● Create program-level learning objectives for professional certificate courses. The number of objectives will depend on the level - beginner, intermediate, or advanced - and the type of certification course.
● Create course-level learning objectives aligned with the overall certification goal.
● Create module-level learning objectives based on skill development relevant to the TG’s career track.
● Review/create Course Outlines for each of the courses.
● Review video scripts and confirm technical accuracy of the content, suggest edits and updates as required. Re-write content and codes as needed. Incorporate one round of internal and client feedback.
● Record talking head videos (onsite/virtually on Zoom) for each course. Incorporate one round of internal and client feedback.
● Provide relevant recorded demos/ screencasts to be integrated in the videos. Check the codes and technical accuracy before providing the demos for integration. Incorporate one round of internal and client feedback.
● For AI/software/tool-based courses, suggest relevant freeware. Write/review and test the codes to check.
● Create/review 2-3 readings per lesson (why and what, 1500 words maximum per reading). The How readings should have detailed instructions/screenshots with short code block type practice that learners can do in their local environment.
● Create One Coach item per lesson - review/reflect on key ideas
● Create/review an ungraded lab per lesson - in-depth activity to apply skills in the learner's local environment.
● Create/review practice quizzes for each lesson and suggest suitable edits, confirm technical accuracy. Incorporate one round of internal and client feedback.
● Create module-level and course-level graded assignments that meet ACE recommendation requirements with 2 additional variations to each item in an assessment bank for each course.
● Create hands-on activities (3-4 lab or any other client preferred format) per course. Incorporate one round of internal and client feedback.
● Create a minimum of one 3-5 min career resources video per course that showcases career path planning.
● For all reviews - validate the content accuracy and provide recommendations/suggestions, write/re-write to fill content gaps as necessary, write/test codes and labs, incorporate 1 round of internal feedback and 2 rounds of client feedback.
● Be available for client discussions and content discussions as and when required.
Requirements:
● 8+ years of experience in data engineering, big data architecture, or analytics roles.
● Strong expertise in Hadoop ecosystem (HDFS, Hive, Pig, HBase) and Apache Spark.
● Proficiency in data integration tools and frameworks like Apache NiFi, Airflow, or Talend. ● Experience with cloud platforms (AWS Redshift, Azure Synapse, Google BigQuery) and data lake/storage solutions.
● Hands-on experience with SQL, Python, Scala, or Java.
● Solid understanding of data warehousing, data modeling, and real-time data streaming (e.g., Kafka, Flink).
● Familiarity with BI tools like Power BI, Tableau, or Looker.
● Strong problem-solving and communication skills with the ability to explain technical concepts to non-technical stakeholders. Preferred Qualifications:
● Master's or Bachelor's degree in Computer Science, Data Science, Engineering, or related field.
● Experience working in regulated industries (e.g., finance, healthcare) with a focus on data compliance and privacy.
● Familiarity with AI/ML frameworks like TensorFlow, PyTorch, or MLlib.
● Certifications in cloud platforms or big data technologies (e.g., AWS Big Data Specialty, GCP Data Engineer).
Timelines and Payout:
Project start date : Immediate Project
Duration: 6 Months
Time Availability : 15-30 hours per course
Job Type : Contract
Work Location : Remote
Job Types: Part-time, Contract, Temporary
Pay: $1,800.00 per month
Application Question(s):
• Are you a US citizen?
• How many years experience do you have with Big Data?
• Do you have teaching experience?
• Do you have video editing experience? Are you camera friendly?
Work Location: Remote