You are viewing this page in an unsupported browser. Please open this link in a browser such as Chrome, Firefox, Safari or Android browser.

The next CutShort event, {{next_cs_event.name}}, in partnership with UpGrad, will happen on {{next_cs_event.startDate | date: 'd MMMM'}}The next CutShort event, {{next_cs_event.name}}, in partnership with UpGrad, will begin in a few hoursThe CutShort event, {{next_cs_event.name}}, in partnership with UpGrad, is LIVE.Join now!

{{hours_remaining}}:{{minutes_remaining}}:{{seconds_remaining}}Want to save 90% of your recruiting time? Learn how in our next webinar on 22nd March at 3 pmLearn more

JOB DESCRIPTION:
We are looking for a Data Engineer with a solid background in scalable systems to work with our engineering team to improve and optimize our platform. You will have significant input into the team’s architectural approach and execution. We are looking for a hands-on programmer who enjoys designing and optimizing data pipelines for large-scale data.
This is NOT a "data scientist" role, so please don't apply if you're looking for that.
RESPONSIBILITIES:
1. Build, maintain and test, performant, scalable data pipelines
2. Work with data scientists and application developers to implement scalable pipelines for data ingest, processing, machine learning and visualization
3. Building interfaces for ingest across various data stores
MUST-HAVE:
1. A track record of building and deploying data pipelines as a part of work or side projects
2. Ability to work with RDBMS, MySQL or Postgres
3. Ability to deploy over cloud infrastructure, at least AWS
4. Demonstrated ability and hunger to learn
GOOD-TO-HAVE:
1. Computer Science degree
2. Expertise in at least one of: Python, Java, Scala
3. Expertise and experience in deploying solutions based on Spark and Kafka
4. Knowledge of container systems like Docker or Kubernetes
5. Experience with NoSQL / graph databases:
6. Knowledge of Machine Learning
Kindly apply only if you are skilled in building data pipelines.

As a Big Data Administrator, you’ll be responsible for the administration and governance of a complex analytics platform that is already changing the way large industrial companies manage their assets. A Big Data Administrator understands cutting-edge tools and frameworks, and is able to determine what the best tools are for any given task. You will enable and work with our other developers to use cutting-edge technologies in the fields of distributed systems, data ingestion and mapping, and machine learning, to name a few. We also strongly encourage everyone to tinker with existing tools, and to stay up to date and test new technologies—all with the aim of ensuring that our existing systems don’t stagnate or deteriorate.
Responsibilities:
As a Big Data Engineer, your responsibilities may include, but are not limited to, the following:
● Build a scalable Big Data Platform designed to serve many different use-cases and requirements
● Build a highly scalable framework for ingesting, transforming and enhancing data at web scale
● Develop data structures and processes using components of the Hadoop ecosystem such as Avro, Hive, Parquet, Impala, Hbase, Kudu, Tez, etc.
● Establish automated build and deployment pipelines
● Implement machine learning models that enable customers to glean hidden insights about their data
● Implementing security and integrating with components such as LDAP, AD, Sentry, Kerberos.
● Strong understanding of row level and role based security concepts such as inheritance
● Establishing scalability benchmarks for predictable scalability thresholds.
Qualifications:
● Bachelor's degree in Computer Science or related field
● 6+ years of system building experience
● 4+ years of programming experience using JVM based languages
● A passion for DevOps and an appreciation for continuous integration/deployment
● A passion for QA and an understanding that testing is not someone else’s responsibility
● Experience automating infrastructure and build processes
● Outstanding programming and problem solving skills
● Strong passion for technology and building great systems
● Excellent communication skills and ability to work using Agile methodologies
● Ability to work quickly and collaboratively in a fast-paced, entrepreneurial environment
● Experience with service-oriented (SOA) and event-driven (EDA) architectures
● Experience using big data solutions in an AWS environment
● Experience with noSQL data stores: Cassandra, HDFS and/or Elasticsearch
● Experience with javascript or associated frameworks
Preferred skills:
We value these qualities, but they’re not required for this role:
● Masters or Ph.D. in related field
● Experience as an open source contributor
● Experience with Akka, stream processing technologies and concurrency frameworks
● Experience with Data modeling
● Experience with Chef, Puppet, Ansible, Salt or equivalent
● Experience with Docker, Mesos and Marathon
● Experience with distributed messaging services, preferably Kafka
● Experience with distributed data processors, preferably Spark
● Experience with Angular, React, Redux, Immutable.js, Rx.js, Node.js or equivalent
● Experience with Reactive and/or Functional programming
● Understanding of Thrift, Avro or protocol buffers

Description Deep experience and understanding of Apache Hadoop and surrounding technologies required; Experience with Spark, Impala, Hive, Flume, Parquet and MapReduce. Strong understanding of development languages to include: Java, Python, Scala, Shell Scripting Expertise in Apache Spark 2. x framework principals and usages. Should be proficient in developing Spark Batch and Streaming job in Python, Scala or Java. Should have proven experience in performance tuning of Spark applications both from application code and configuration perspective. Should be proficient in Kafka and integration with Spark. Should be proficient in Spark SQL and data warehousing techniques using Hive. Should be very proficient in Unix shell scripting and in operating on Linux. Should have knowledge about any cloud based infrastructure. Good experience in tuning Spark applications and performance improvements. Strong understanding of data profiling concepts and ability to operationalize analyses into design and development activities Experience with best practices of software development; Version control systems, automated builds, etc. Experienced in and able to lead the following phases of the Software Development Life Cycle on any project (feasibility planning, analysis, development, integration, test and implementation) Capable of working within the team or as an individual Experience to create technical documentation

Description Does solving complex business problems and real world challenges interest you Do you enjoy seeing the impact your contributions make on a daily basis Are you passionate about using data analytics to provide game changing solutions to the Global 2000 clients Do you thrive in a dynamic work environment that constantly pushes you to be the best you can be and more Are you ready to work with smart colleagues who drive for excellence in everything they do If you possess a solutions mindset, strong analytical skills, and commitment to be part of a tremendous journey, come join our growing, global team. See what Saama can do for your career and for your journey. Position: Java/ Big Data Lead (2162) Location: Hinjewadi Phase 1, Pune Type: Permanent Full time Requirements: Candidate should be able - Define application level architecture and guide low level of Database design Gather technical requirements and propose solutions based on client s business and architectural needs Interact with prospective customers during product demos/ evaluations Internally work with technology and business groups to define project specifications Showcase experience on cloud based implementations and technically manage Bigdata and j2EE projects Showcase experience hands-on programming and debugging skills on Spring, Hibernate, Java, JavaScript, JSP/ Servlet, J2EE design patterns / Python Have knowledge on service Integration Concepts (especially with RESTFUL services/ SOAP based web services) Design and develop solutions for Non-Functional Requirements (Performance analysis & tuning, Benchmarking/ load testing, Security) Impact on the business: Plays an important role in making Saama s Solutions game changers for our strategic partners by using data science to solve core, complex business challenges. Key relationships: Sales & pre-sales Product management Engineering Client organization: account management & delivery Saama Competencies: INTEGRITY: we do the right things. INNOVATION: we change the game. TRANSPARENCY: we communicate openly COLLABORATION: we work as one team PROBLEM-SOLVING: we solve core, complex business challenges ENJOY & CELEBRATE: we have fun Competencies: Self-starter who gets results with minimal support and direction in a fast-paced environment. Takes initiative; challenges the status quo to drive change. Learns quickly; takes smart risks to experiment and learn. Works well with others; builds trust and maintains credibility. Planful: identifies and confirms key requirements in dynamic environments; anticipates tasks and contingencies. Communicates effectively; productive communication with clients and all key stakeholders communication in both verbal and written communication. Stays the course despite challenges & setbacks. Works well under pressure. Strong analytical skills; able to apply inductive and deductive thinking to generate solutions for complex problems

We are a start-up in India seeking excellence in everything we do with an unwavering curiosity and enthusiasm. We build simplified new-age AI driven Big Data Analytics platform for Global Enterprises and solve their biggest business challenges. Our Engineers develop fresh intuitive solutions keeping the user in the center of everything.
As a Cloud-ML Engineer, you will design and implement ML solutions for customer use cases and problem solve complex technical customer challenges.
Expectations and Tasks
- Total of 7+ years of experience with minimum of 2 years in Hadoop technologies like HDFS, Hive, MapReduce
- Experience working with recommendation engines, data pipelines, or distributed machine learning and experience with data analytics and data visualization techniques and software.
- Experience with core Data Science techniques such as regression, classification or clustering, and experience with deep learning frameworks
- Experience in NLP, R and Python
- Experience in performance tuning and optimization techniques to process big data from heterogeneous sources.
- Ability to communicate clearly and concisely across technology and the business teams.
- Excellent Problem solving and Technical troubleshooting skills.
- Ability to handle multiple projects and prioritize tasks in a rapidly changing environment.
Technical Skills
Core Java, Multithreading, Collections, OOPS, Python, R, Apache Spark, MapReduce, Hive, HDFS, Hadoop, MongoDB, Scala
We are a retained Search Firm employed by our client - Technology Start-up @ Bangalore. Interested candidates can share their resumes with me - Jia@TalentSculpt.com. I will respond to you within 24 hours. Online assessments and pre-employment screening are part of the selection process.

We are a US based startup on a mission to personalize learning for everyone and also make it affordable. We are currently building the most advanced adaptive engine that applies learning methodologies based on the cognition and retention of each student.
Our product vision is to enable students improve their learning outcomes in the shortest possible time to enable them to gain mastery, better grades and ultimately have greater opportunities in their desired field.
Our solutions are currently being used at various educational institutions to help students overcome hurdles during their learning process.
If you are passionate about our mission, come and talk to us.
We provide an open collaborative environment, a great team and a fun place to work along with excellent benefits and attractive stock options.